Skip to content

Commit 91ce496

Browse files
pre-commit-ci[bot]VeckoTheGecko
authored andcommitted
[pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
1 parent 3fe2ad3 commit 91ce496

File tree

8 files changed

+18
-21
lines changed

8 files changed

+18
-21
lines changed

src/parcels/_core/particlefile.py

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@
99
import cftime
1010
import numpy as np
1111
import pyarrow as pa
12-
from parcels._typing import PathLike
1312
import pyarrow.parquet as pq
1413

1514
import parcels
1615
from parcels._core.particle import ParticleClass
1716
from parcels._core.utils.time import timedelta_to_float
1817
from parcels._reprs import particlefile_repr
18+
from parcels._typing import PathLike
1919

2020
if TYPE_CHECKING:
2121
from parcels._core.particle import Variable
@@ -69,14 +69,16 @@ def __init__(self, path: PathLike, outputdt):
6969
path = Path(path)
7070

7171
if path.suffix != ".parquet":
72-
raise ValueError(f"ParticleFile data is stored in Parquet files - file extension must be '.parquet'. Got {path.suffix=!r}.")
72+
raise ValueError(
73+
f"ParticleFile data is stored in Parquet files - file extension must be '.parquet'. Got {path.suffix=!r}."
74+
)
7375

7476
if outputdt <= 0:
7577
raise ValueError(f"outputdt must be positive/non-zero. Got {outputdt=!r}")
7678

7779
self._outputdt = outputdt
7880

79-
self._path = path # TODO v4: Consider https://arrow.apache.org/docs/python/getstarted.html#working-with-large-data - though a significant question becomes how to partition, perhaps using a particle variable "partition"?
81+
self._path = path # TODO v4: Consider https://arrow.apache.org/docs/python/getstarted.html#working-with-large-data - though a significant question becomes how to partition, perhaps using a particle variable "partition"?
8082
self._writer: pq.ParquetWriter | None = None
8183
if path.exists():
8284
# TODO: Add logic for recovering/appending to existing parquet file
@@ -138,7 +140,7 @@ def write(self, pset: ParticleSet, time, indices=None):
138140
indices_to_write = _to_write_particles(particle_data, time)
139141
else:
140142
indices_to_write = indices
141-
143+
142144
self._writer.write_table(
143145
pa.table({v.name: pa.array(particle_data[v.name][indices_to_write]) for v in vars_to_write}),
144146
)

src/parcels/_core/particleset.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -453,7 +453,7 @@ def execute(
453453

454454
if output_file is not None:
455455
output_file.close()
456-
456+
457457
if verbose_progress:
458458
pbar.close()
459459

src/parcels/_reprs.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,6 @@
77

88
import numpy as np
99
import xarray as xr
10-
from zarr.storage import DirectoryStore
1110

1211
if TYPE_CHECKING:
1312
from parcels import Field, FieldSet, ParticleSet
@@ -178,6 +177,5 @@ def _format_list_items_multiline(items: list[str] | dict, level: int = 1, with_b
178177
return "\n".join([textwrap.indent(e, indentation_str) for e in entries])
179178

180179

181-
182180
def is_builtin_object(obj):
183181
return obj.__class__.__module__ == "builtins"

tests/conftest.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ def tmp_parquet(tmp_path, request):
1212
def tmp_store():
1313
return MemoryStore()
1414

15+
1516
@pytest.fixture
1617
def tmp_parquet(tmp_path):
17-
return tmp_path / 'tmp.parquet'
18-
18+
return tmp_path / "tmp.parquet"

tests/test_advection.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
import numpy as np
2+
import pandas as pd
23
import pytest
34
import xarray as xr
45

56
import parcels
67
import parcels.tutorial
7-
import pandas as pd
88
from parcels import (
99
Field,
1010
FieldSet,

tests/test_fieldset.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,8 +3,8 @@
33
import cf_xarray # noqa: F401
44
import cftime
55
import numpy as np
6-
import pytest
76
import pandas as pd
7+
import pytest
88
import xarray as xr
99

1010
from parcels import Field, ParticleFile, ParticleSet, VectorField, XGrid, convert
@@ -96,7 +96,6 @@ def test_fieldset_gridset(fieldset):
9696
assert len(fieldset.gridset) == 2
9797

9898

99-
10099
def test_fieldset_no_UV(tmp_parquet):
101100
grid = XGrid.from_dataset(ds, mesh="flat")
102101
fieldset = FieldSet([Field("P", ds["U_A_grid"], grid, interp_method=XLinear)])

tests/test_particlefile.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -2,13 +2,13 @@
22
import tempfile
33
from contextlib import nullcontext as does_not_raise
44
from datetime import datetime, timedelta
5-
import pyarrow.parquet as pq
6-
import pandas as pd
5+
76
import numpy as np
7+
import pandas as pd
88
import pyarrow as pa
9+
import pyarrow.parquet as pq
910
import pytest
1011
import xarray as xr
11-
from zarr.storage import MemoryStore
1212

1313
import parcels.tutorial
1414
from parcels import (
@@ -22,9 +22,9 @@
2222
VectorField,
2323
XGrid,
2424
)
25-
from parcels._core.particle import Particle, create_particle_data, get_default_particle
25+
from parcels._core.particle import Particle, get_default_particle
2626
from parcels._core.particlefile import _get_schema
27-
from parcels._core.utils.time import TimeInterval, timedelta_to_float
27+
from parcels._core.utils.time import timedelta_to_float
2828
from parcels._datasets.structured.generated import peninsula_dataset
2929
from parcels._datasets.structured.generic import datasets
3030
from parcels.convert import copernicusmarine_to_sgrid
@@ -343,7 +343,7 @@ def Update_lon(particles, fieldset): # pragma: no cover
343343
pset.execute(Update_lon, runtime=np.timedelta64(11, "s"), dt=np.timedelta64(2, "s"), output_file=ofile)
344344

345345
df = pd.read_parquet(tmp_parquet)
346-
assert np.allclose(df['lon'].values, [0, 3, 6, 9])
346+
assert np.allclose(df["lon"].values, [0, 3, 6, 9])
347347
pytest.skip("# TODO: Need to figure out how times work with parquet output (#2386)")
348348
assert np.allclose(timedelta_to_float(df.time.values - df.time.values[0, 0]), [0, 3, 6, 9])
349349

tests/test_uxadvection.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
import numpy as np
2+
import pandas as pd
23
import pytest
3-
import xarray as xr
44

55
import parcels
66
from parcels._datasets.unstructured.generic import datasets as datasets_unstructured
@@ -9,8 +9,6 @@
99
AdvectionRK2,
1010
AdvectionRK4,
1111
)
12-
import pandas as pd
13-
1412

1513

1614
@pytest.mark.parametrize("integrator", [AdvectionEE, AdvectionRK2, AdvectionRK4])

0 commit comments

Comments
 (0)