@@ -31,9 +31,9 @@ async def _decode_single(
3131 cdata = chunk_bytes .as_array_like ()
3232 # decompress
3333 if self .compressor :
34- chunk = await asyncio .to_thread (self .compressor .decode , cdata ) # type: ignore[arg-type]
34+ chunk = await asyncio .to_thread (self .compressor .decode , cdata )
3535 else :
36- chunk = cdata # type: ignore[assignment]
36+ chunk = cdata
3737
3838 # apply filters
3939 if self .filters :
@@ -54,7 +54,7 @@ async def _decode_single(
5454 # is an object array. In this case, we need to convert the object
5555 # array to the correct dtype.
5656
57- chunk = np .array (chunk ).astype (chunk_spec .dtype .to_native_dtype ()) # type: ignore[assignment]
57+ chunk = np .array (chunk ).astype (chunk_spec .dtype .to_native_dtype ())
5858
5959 elif chunk .dtype != object :
6060 # If we end up here, someone must have hacked around with the filters.
@@ -83,18 +83,16 @@ async def _encode_single(
8383 # apply filters
8484 if self .filters :
8585 for f in self .filters :
86- chunk = await asyncio .to_thread (f .encode , chunk ) # type: ignore[arg-type]
87-
86+ chunk = await asyncio .to_thread (f .encode , chunk )
8887 # check object encoding
8988 if ensure_ndarray_like (chunk ).dtype == object :
9089 raise RuntimeError ("cannot write object array without object codec" )
9190
9291 # compress
9392 if self .compressor :
94- cdata = await asyncio .to_thread (self .compressor .encode , chunk ) # type: ignore[arg-type]
93+ cdata = await asyncio .to_thread (self .compressor .encode , chunk )
9594 else :
96- cdata = chunk # type: ignore[assignment]
97-
95+ cdata = chunk
9896 cdata = ensure_bytes (cdata )
9997 return chunk_spec .prototype .buffer .from_bytes (cdata )
10098
0 commit comments