Skip to content

Commit ccfdd81

Browse files
committed
update bodo and fix test issues
1 parent cae2425 commit ccfdd81

File tree

5 files changed

+59
-48
lines changed

5 files changed

+59
-48
lines changed

poetry.lock

Lines changed: 38 additions & 38 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -79,7 +79,7 @@ huggingface-hub = { version = ">=0.24.0", optional = true }
7979
psycopg2-binary = { version = ">=2.9.6", optional = true }
8080
sqlalchemy = { version = "^2.0.18", optional = true }
8181
getdaft = { version = ">=0.2.12", optional = true }
82-
bodo = { version = ">=2025.7.3", python = ">=3.9,<3.14", optional = true }
82+
bodo = { version = ">=2025.7.4", optional = true }
8383
cachetools = ">=5.5,<7.0"
8484
pyiceberg-core = { version = "^0.5.1", optional = true }
8585
polars = { version = "^1.21.0", optional = true }

tests/integration/test_reads.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -346,6 +346,7 @@ def test_daft_nan_rewritten(catalog: Catalog) -> None:
346346

347347

348348
@pytest.mark.integration
349+
@pytest.mark.filterwarnings("ignore")
349350
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")])
350351
def test_bodo_nan(catalog: Catalog) -> None:
351352
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten")

tests/integration/test_writes/test_partitioned_writes.py

Lines changed: 9 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -451,6 +451,11 @@ def test_dynamic_partition_overwrite_unpartitioned_evolve_to_identity_transform(
451451

452452
@pytest.mark.integration
453453
def test_summaries_with_null(spark: SparkSession, session_catalog: Catalog, arrow_table_with_null: pa.Table) -> None:
454+
import pyarrow
455+
from packaging import version
456+
457+
under_20_arrow = version.parse(pyarrow.__version__) < version.parse("20.0.0")
458+
454459
identifier = "default.arrow_table_summaries"
455460

456461
try:
@@ -547,14 +552,14 @@ def test_summaries_with_null(spark: SparkSession, session_catalog: Catalog, arro
547552
"total-records": "6",
548553
}
549554
assert summaries[5] == {
550-
"removed-files-size": "16174",
555+
"removed-files-size": "15774" if under_20_arrow else "16174",
551556
"changed-partition-count": "2",
552557
"total-equality-deletes": "0",
553558
"deleted-data-files": "4",
554559
"total-position-deletes": "0",
555560
"total-delete-files": "0",
556561
"deleted-records": "4",
557-
"total-files-size": "8884",
562+
"total-files-size": "8684" if under_20_arrow else "8884",
558563
"total-data-files": "2",
559564
"total-records": "2",
560565
}
@@ -564,9 +569,9 @@ def test_summaries_with_null(spark: SparkSession, session_catalog: Catalog, arro
564569
"total-equality-deletes": "0",
565570
"added-records": "2",
566571
"total-position-deletes": "0",
567-
"added-files-size": "8087",
572+
"added-files-size": "7887" if under_20_arrow else "8087",
568573
"total-delete-files": "0",
569-
"total-files-size": "16971",
574+
"total-files-size": "16571" if under_20_arrow else "16971",
570575
"total-data-files": "4",
571576
"total-records": "4",
572577
}

tests/integration/test_writes/test_writes.py

Lines changed: 10 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -269,6 +269,11 @@ def test_summaries(spark: SparkSession, session_catalog: Catalog, arrow_table_wi
269269

270270
@pytest.mark.integration
271271
def test_summaries_partial_overwrite(spark: SparkSession, session_catalog: Catalog) -> None:
272+
import pyarrow
273+
from packaging import version
274+
275+
under_20_arrow = version.parse(pyarrow.__version__) < version.parse("20.0.0")
276+
272277
identifier = "default.test_summaries_partial_overwrite"
273278
TEST_DATA = {
274279
"id": [1, 2, 3, 1, 1],
@@ -309,13 +314,13 @@ def test_summaries_partial_overwrite(spark: SparkSession, session_catalog: Catal
309314
# APPEND
310315
assert summaries[0] == {
311316
"added-data-files": "3",
312-
"added-files-size": "2618",
317+
"added-files-size": "2570" if under_20_arrow else "2618",
313318
"added-records": "5",
314319
"changed-partition-count": "3",
315320
"total-data-files": "3",
316321
"total-delete-files": "0",
317322
"total-equality-deletes": "0",
318-
"total-files-size": "2618",
323+
"total-files-size": "2570" if under_20_arrow else "2618",
319324
"total-position-deletes": "0",
320325
"total-records": "5",
321326
}
@@ -344,16 +349,16 @@ def test_summaries_partial_overwrite(spark: SparkSession, session_catalog: Catal
344349
assert len(files) == 3
345350
assert summaries[1] == {
346351
"added-data-files": "1",
347-
"added-files-size": "875",
352+
"added-files-size": "859" if under_20_arrow else "875",
348353
"added-records": "2",
349354
"changed-partition-count": "1",
350355
"deleted-data-files": "1",
351356
"deleted-records": "3",
352-
"removed-files-size": "882",
357+
"removed-files-size": "866" if under_20_arrow else "882",
353358
"total-data-files": "3",
354359
"total-delete-files": "0",
355360
"total-equality-deletes": "0",
356-
"total-files-size": "2611",
361+
"total-files-size": "2563" if under_20_arrow else "2611",
357362
"total-position-deletes": "0",
358363
"total-records": "4",
359364
}

0 commit comments

Comments
 (0)