Skip to content

Commit afe8df8

Browse files
author
ShreyeshArangath
committed
attempt to fix lint
1 parent a0ddc25 commit afe8df8

File tree

3 files changed

+17
-26
lines changed

3 files changed

+17
-26
lines changed

crates/core/src/dataframe.rs

Lines changed: 11 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -41,6 +41,11 @@ use datafusion::execution::context::TaskContext;
4141
use datafusion::logical_expr::SortExpr;
4242
use datafusion::logical_expr::dml::InsertOp;
4343
use datafusion::parquet::basic::{BrotliLevel, Compression, GzipLevel, ZstdLevel};
44+
use datafusion::physical_plan::{
45+
ExecutionPlan as DFExecutionPlan, collect as df_collect,
46+
collect_partitioned as df_collect_partitioned, execute_stream as df_execute_stream,
47+
execute_stream_partitioned as df_execute_stream_partitioned,
48+
};
4449
use datafusion::prelude::*;
4550
use datafusion_python_util::{is_ipython_env, spawn_future, wait_for_future};
4651
use futures::{StreamExt, TryStreamExt};
@@ -52,13 +57,6 @@ use pyo3::pybacked::PyBackedStr;
5257
use pyo3::types::{PyCapsule, PyList, PyTuple, PyTupleMethods};
5358

5459
use crate::common::data_type::PyScalarValue;
55-
use datafusion::physical_plan::{
56-
ExecutionPlan as DFExecutionPlan,
57-
collect as df_collect,
58-
collect_partitioned as df_collect_partitioned,
59-
execute_stream as df_execute_stream,
60-
execute_stream_partitioned as df_execute_stream_partitioned,
61-
};
6260
use crate::errors::{PyDataFusionError, PyDataFusionResult, py_datafusion_err};
6361
use crate::expr::PyExpr;
6462
use crate::expr::sort_expr::{PySortExpr, to_sort_expressions};
@@ -672,8 +670,8 @@ impl PyDataFrame {
672670
/// guarantee of the order of the result.
673671
fn collect<'py>(&self, py: Python<'py>) -> PyResult<Vec<Bound<'py, PyAny>>> {
674672
let (plan, task_ctx) = self.create_and_cache_plan(py)?;
675-
let batches = wait_for_future(py, df_collect(plan, task_ctx))?
676-
.map_err(PyDataFusionError::from)?;
673+
let batches =
674+
wait_for_future(py, df_collect(plan, task_ctx))?.map_err(PyDataFusionError::from)?;
677675
// cannot use PyResult<Vec<RecordBatch>> return type due to
678676
// https://github.com/PyO3/pyo3/issues/1813
679677
batches.into_iter().map(|rb| rb.to_pyarrow(py)).collect()
@@ -1187,9 +1185,10 @@ impl PyDataFrame {
11871185

11881186
fn execute_stream_partitioned(&self, py: Python) -> PyResult<Vec<PyRecordBatchStream>> {
11891187
let (plan, task_ctx) = self.create_and_cache_plan(py)?;
1190-
let streams = spawn_future(py, async move {
1191-
df_execute_stream_partitioned(plan, task_ctx)
1192-
})?;
1188+
let streams = spawn_future(
1189+
py,
1190+
async move { df_execute_stream_partitioned(plan, task_ctx) },
1191+
)?;
11931192
Ok(streams.into_iter().map(PyRecordBatchStream::new).collect())
11941193
}
11951194

crates/core/src/metrics.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
use std::collections::HashMap;
1919
use std::sync::Arc;
2020

21-
use datafusion::physical_plan::metrics::{MetricValue, MetricsSet, Metric, Timestamp};
21+
use datafusion::physical_plan::metrics::{Metric, MetricValue, MetricsSet, Timestamp};
2222
use pyo3::prelude::*;
2323

2424
#[pyclass(frozen, name = "MetricsSet", module = "datafusion")]

python/tests/test_plans.py

Lines changed: 5 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ def test_metric_properties() -> None:
9292
assert len(metric.name) > 0
9393
assert metric.partition is None or isinstance(metric.partition, int)
9494
assert metric.value is None or isinstance(
95-
metric.value, (int, datetime.datetime)
95+
metric.value, int | datetime.datetime
9696
)
9797
assert isinstance(metric.labels(), dict)
9898
mr = repr(metric)
@@ -136,9 +136,7 @@ def test_collect_partitioned_metrics() -> None:
136136
plan = df.execution_plan()
137137

138138
output_rows_values = [
139-
ms.output_rows
140-
for _, ms in plan.collect_metrics()
141-
if ms.output_rows is not None
139+
ms.output_rows for _, ms in plan.collect_metrics() if ms.output_rows is not None
142140
]
143141
assert 2 in output_rows_values, f"Expected 2 in {output_rows_values}"
144142

@@ -153,9 +151,7 @@ def test_execute_stream_metrics() -> None:
153151

154152
plan = df.execution_plan()
155153
output_rows_values = [
156-
ms.output_rows
157-
for _, ms in plan.collect_metrics()
158-
if ms.output_rows is not None
154+
ms.output_rows for _, ms in plan.collect_metrics() if ms.output_rows is not None
159155
]
160156
assert 2 in output_rows_values, f"Expected 2 in {output_rows_values}"
161157

@@ -171,9 +167,7 @@ def test_execute_stream_partitioned_metrics() -> None:
171167

172168
plan = df.execution_plan()
173169
output_rows_values = [
174-
ms.output_rows
175-
for _, ms in plan.collect_metrics()
176-
if ms.output_rows is not None
170+
ms.output_rows for _, ms in plan.collect_metrics() if ms.output_rows is not None
177171
]
178172
assert 2 in output_rows_values, f"Expected 2 in {output_rows_values}"
179173

@@ -233,8 +227,6 @@ def test_collect_twice_has_metrics() -> None:
233227

234228
plan = df.execution_plan()
235229
output_rows_values = [
236-
ms.output_rows
237-
for _, ms in plan.collect_metrics()
238-
if ms.output_rows is not None
230+
ms.output_rows for _, ms in plan.collect_metrics() if ms.output_rows is not None
239231
]
240232
assert len(output_rows_values) > 0

0 commit comments

Comments
 (0)