@@ -1167,7 +1167,7 @@ async def test_crawler_multiple_stops_in_parallel() -> None:
11671167 # Set concurrency to 2 to ensure two urls are being visited in parallel.
11681168 crawler = BasicCrawler (concurrency_settings = ConcurrencySettings (desired_concurrency = 2 , max_concurrency = 2 ))
11691169
1170- both_handlers_started = asyncio .Barrier (2 ) # type :ignore[attr-defined ] # Test is skipped in older Python versions.
1170+ both_handlers_started = asyncio .Barrier (2 ) # ty :ignore[unresolved-attribute ] # Test is skipped in older Python versions.
11711171 only_one_handler_at_a_time = asyncio .Semaphore (1 )
11721172
11731173 @crawler .router .default_handler
@@ -1351,7 +1351,7 @@ async def test_context_use_state_race_condition_in_handlers(key_value_store: Key
13511351 Result should be incremented by 2.
13521352 Method `use_state` must be implemented in a way that prevents race conditions in such scenario."""
13531353 # Test is skipped in older Python versions.
1354- from asyncio import Barrier # type :ignore[attr-defined ] # noqa: PLC0415
1354+ from asyncio import Barrier # ty :ignore[unresolved-import ] # noqa: PLC0415
13551355
13561356 crawler = BasicCrawler ()
13571357 store = await crawler .get_key_value_store ()
@@ -1392,7 +1392,7 @@ async def test_timeout_in_handler(sleep_type: str) -> None:
13921392 Crawler should attempt to retry it.
13931393 This test creates situation where the request handler times out twice, on third retry it does not time out."""
13941394 # Test is skipped in older Python versions.
1395- from asyncio import timeout # type :ignore[attr-defined ] # noqa: PLC0415
1395+ from asyncio import timeout # ty :ignore[unresolved-import ] # noqa: PLC0415
13961396
13971397 non_realtime_system_coefficient = 10
13981398 handler_timeout = timedelta (seconds = 1 )
0 commit comments