@@ -190,8 +190,9 @@ def test_query_filter_v1_v2_append_null(
190190@pytest .mark .integration
191191def test_summaries (spark : SparkSession , session_catalog : Catalog , arrow_table_with_null : pa .Table ) -> None :
192192 identifier = "default.arrow_table_summaries"
193- tbl = _create_table (session_catalog , identifier , {"format-version" : "1" }, 2 * [arrow_table_with_null ])
194- tbl .overwrite (arrow_table_with_null )
193+ tbl = _create_table (session_catalog , identifier , {"format-version" : "1" }, 2 * [arrow_table_with_null ]) # append + append
194+ tbl .overwrite (arrow_table_with_null ) # delete + append
195+ tbl .delete (delete_filter = "int == 1" ) # overwrite, deletes 1 row
195196
196197 rows = spark .sql (
197198 f"""
@@ -202,14 +203,14 @@ def test_summaries(spark: SparkSession, session_catalog: Catalog, arrow_table_wi
202203 ).collect ()
203204
204205 operations = [row .operation for row in rows ]
205- assert operations == ["append" , "append" , "delete" , "append" ]
206+ assert operations == ["append" , "append" , "delete" , "append" , "overwrite" ]
206207
207208 summaries = [row .summary for row in rows ]
208209
209210 file_size = int (summaries [0 ]["added-files-size" ])
210211 assert file_size > 0
211212
212- # Append
213+ # Append from _create_table
213214 assert summaries [0 ] == {
214215 "added-data-files" : "1" ,
215216 "added-files-size" : str (file_size ),
@@ -222,7 +223,7 @@ def test_summaries(spark: SparkSession, session_catalog: Catalog, arrow_table_wi
222223 "total-records" : "3" ,
223224 }
224225
225- # Append
226+ # Append from _create_table
226227 assert summaries [1 ] == {
227228 "added-data-files" : "1" ,
228229 "added-files-size" : str (file_size ),
@@ -235,7 +236,7 @@ def test_summaries(spark: SparkSession, session_catalog: Catalog, arrow_table_wi
235236 "total-records" : "6" ,
236237 }
237238
238- # Delete
239+ # Delete from tbl.overwrite
239240 assert summaries [2 ] == {
240241 "deleted-data-files" : "2" ,
241242 "deleted-records" : "6" ,
@@ -248,7 +249,7 @@ def test_summaries(spark: SparkSession, session_catalog: Catalog, arrow_table_wi
248249 "total-records" : "0" ,
249250 }
250251
251- # Overwrite
252+ # Append from tbl.overwrite
252253 assert summaries [3 ] == {
253254 "added-data-files" : "1" ,
254255 "added-files-size" : str (file_size ),
@@ -261,6 +262,22 @@ def test_summaries(spark: SparkSession, session_catalog: Catalog, arrow_table_wi
261262 "total-records" : "3" ,
262263 }
263264
265+ # Delete from tbl.delete
266+ assert summaries [4 ] == {
267+ "added-data-files" : "1" ,
268+ "added-files-size" : "4342" ,
269+ "added-records" : "2" ,
270+ "deleted-data-files" : "1" ,
271+ "deleted-records" : "3" ,
272+ "removed-files-size" : "4406" ,
273+ "total-data-files" : "1" ,
274+ "total-delete-files" : "0" ,
275+ "total-equality-deletes" : "0" ,
276+ "total-files-size" : "4342" ,
277+ "total-position-deletes" : "0" ,
278+ "total-records" : "2" ,
279+ }
280+
264281
265282@pytest .mark .integration
266283def test_data_files (spark : SparkSession , session_catalog : Catalog , arrow_table_with_null : pa .Table ) -> None :
0 commit comments