@@ -273,7 +273,6 @@ def test_data_files(spark: SparkSession, session_catalog: Catalog, arrow_table_w
273273 # Since we don't rewrite, this should produce a new manifest with an ADDED entry
274274 tbl .append (arrow_table_with_null )
275275
276-
277276 rows = spark .sql (
278277 f"""
279278 SELECT added_data_files_count, existing_data_files_count, deleted_data_files_count
@@ -285,27 +284,32 @@ def test_data_files(spark: SparkSession, session_catalog: Catalog, arrow_table_w
285284 assert [row .existing_data_files_count for row in rows ] == [0 , 0 , 0 , 0 , 0 ]
286285 assert [row .deleted_data_files_count for row in rows ] == [0 , 1 , 0 , 0 , 0 ]
287286
288-
287+ tests / integration / test_writes / test_writes . py
289288@pytest .mark .integration
290289@pytest .mark .parametrize ("format_version" , [1 , 2 ])
291- def test_object_storage_data_files (spark : SparkSession , session_catalog : Catalog , arrow_table_with_null : pa .Table , format_version : int ) -> None :
292- # TODO: What to do about "tbl.add_files()"?
293- identifier = "default.object_stored_table"
294-
295- tbl = _create_table (session_catalog , identifier , {"format-version" : format_version , "write.object-storage.enabled" : True }, [])
290+ def test_object_storage_data_files (
291+ spark : SparkSession , session_catalog : Catalog , arrow_table_with_null : pa .Table , format_version : int
292+ ) -> None :
293+ tbl = _create_table (
294+ session_catalog = session_catalog ,
295+ identifier = "default.object_stored" ,
296+ properties = {"format-version" : format_version , "write.object-storage.enabled" : True },
297+ data = [arrow_table_with_null ],
298+ )
296299 tbl .append (arrow_table_with_null )
297300
298- paths = tbl .inspect .entries ().to_pydict ()["data_file" ]
299- assert len (paths ) == 1
300- location = paths [0 ]["file_path" ]
301+ paths = tbl .inspect .data_files ().to_pydict ()["file_path" ]
302+ assert len (paths ) == 2
301303
302- parts = location .split ("/" )
303- assert len (parts ) == 11
304+ for location in paths :
305+ assert location .startswith ("s3://warehouse/default/object_stored/data/" )
306+ parts = location .split ("/" )
307+ assert len (parts ) == 11
304308
305- assert location . startswith ( "s3://warehouse/default/object_stored_table/data/" )
306- for i in range (6 , 10 ):
307- assert len ( parts [ i ]) == ( 8 if i == 9 else 4 )
308- assert all (c in "01" for c in parts [i ])
309+ # Entropy binary directories should have been injected
310+ for i in range (6 , 10 ):
311+ assert parts [ i ]
312+ assert all (c in "01" for c in parts [i ])
309313
310314
311315@pytest .mark .integration
0 commit comments