@@ -19,6 +19,9 @@ def _is_streaming_response(response):
1919
2020
2121# Streaming in chat models
22+ @pytest .mark .skip (
23+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
24+ )
2225@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
2326def test_stream_chat_models (model_name ):
2427 name = f"test_stream_chat_models-{ create_uuid ()} "
@@ -73,6 +76,9 @@ def test_stream_chat_models(model_name):
7376
7477
7578# Streaming in completions models
79+ @pytest .mark .skip (
80+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
81+ )
7682@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
7783def test_stream_completions_models (model_name ):
7884 name = f"test_stream_completions_models-{ create_uuid ()} "
@@ -124,6 +130,9 @@ def test_stream_completions_models(model_name):
124130
125131
126132# Invoke in chat models
133+ @pytest .mark .skip (
134+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
135+ )
127136@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
128137def test_invoke_chat_models (model_name ):
129138 name = f"test_invoke_chat_models-{ create_uuid ()} "
@@ -171,6 +180,9 @@ def test_invoke_chat_models(model_name):
171180
172181
173182# Invoke in completions models
183+ @pytest .mark .skip (
184+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
185+ )
174186@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
175187def test_invoke_in_completions_models (model_name ):
176188 name = f"test_invoke_in_completions_models-{ create_uuid ()} "
@@ -217,6 +229,9 @@ def test_invoke_in_completions_models(model_name):
217229 assert generation .latency is not None
218230
219231
232+ @pytest .mark .skip (
233+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
234+ )
220235@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
221236def test_batch_in_completions_models (model_name ):
222237 name = f"test_batch_in_completions_models-{ create_uuid ()} "
@@ -263,6 +278,9 @@ def test_batch_in_completions_models(model_name):
263278 assert generation .latency is not None
264279
265280
281+ @pytest .mark .skip (
282+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
283+ )
266284@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
267285def test_batch_in_chat_models (model_name ):
268286 name = f"test_batch_in_chat_models-{ create_uuid ()} "
@@ -309,6 +327,9 @@ def test_batch_in_chat_models(model_name):
309327
310328
311329# Async stream in chat models
330+ @pytest .mark .skip (
331+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
332+ )
312333@pytest .mark .asyncio
313334@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
314335async def test_astream_chat_models (model_name ):
@@ -363,6 +384,9 @@ async def test_astream_chat_models(model_name):
363384
364385
365386# Async stream in completions model
387+ @pytest .mark .skip (
388+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
389+ )
366390@pytest .mark .asyncio
367391@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
368392async def test_astream_completions_models (model_name ):
@@ -417,6 +441,9 @@ async def test_astream_completions_models(model_name):
417441
418442
419443# Async invoke in chat models
444+ @pytest .mark .skip (
445+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
446+ )
420447@pytest .mark .asyncio
421448@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
422449async def test_ainvoke_chat_models (model_name ):
@@ -465,6 +492,9 @@ async def test_ainvoke_chat_models(model_name):
465492 assert generation .latency is not None
466493
467494
495+ @pytest .mark .skip (
496+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
497+ )
468498@pytest .mark .asyncio
469499@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
470500async def test_ainvoke_in_completions_models (model_name ):
@@ -516,6 +546,9 @@ async def test_ainvoke_in_completions_models(model_name):
516546
517547
518548# Sync batch in chains and chat models
549+ @pytest .mark .skip (
550+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
551+ )
519552@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
520553def test_chains_batch_in_chat_models (model_name ):
521554 name = f"test_chains_batch_in_chat_models-{ create_uuid ()} "
@@ -564,6 +597,9 @@ def test_chains_batch_in_chat_models(model_name):
564597 assert generation .latency is not None
565598
566599
600+ @pytest .mark .skip (
601+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
602+ )
567603@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
568604def test_chains_batch_in_completions_models (model_name ):
569605 name = f"test_chains_batch_in_completions_models-{ create_uuid ()} "
@@ -613,6 +649,9 @@ def test_chains_batch_in_completions_models(model_name):
613649
614650
615651# Async batch call with chains and chat models
652+ @pytest .mark .skip (
653+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
654+ )
616655@pytest .mark .asyncio
617656@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
618657async def test_chains_abatch_in_chat_models (model_name ):
@@ -663,6 +702,9 @@ async def test_chains_abatch_in_chat_models(model_name):
663702
664703
665704# Async batch call with chains and completions models
705+ @pytest .mark .skip (
706+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
707+ )
666708@pytest .mark .asyncio
667709@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
668710async def test_chains_abatch_in_completions_models (model_name ):
@@ -709,6 +751,9 @@ async def test_chains_abatch_in_completions_models(model_name):
709751
710752
711753# Async invoke in chains and chat models
754+ @pytest .mark .skip (
755+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
756+ )
712757@pytest .mark .asyncio
713758@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" ])
714759async def test_chains_ainvoke_chat_models (model_name ):
@@ -762,6 +807,9 @@ async def test_chains_ainvoke_chat_models(model_name):
762807
763808
764809# Async invoke in chains and completions models
810+ @pytest .mark .skip (
811+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
812+ )
765813@pytest .mark .asyncio
766814@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
767815async def test_chains_ainvoke_completions_models (model_name ):
@@ -813,6 +861,9 @@ async def test_chains_ainvoke_completions_models(model_name):
813861
814862
815863# Async streaming in chat models
864+ @pytest .mark .skip (
865+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
866+ )
816867@pytest .mark .asyncio
817868@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo" , "gpt-4" ])
818869async def test_chains_astream_chat_models (model_name ):
@@ -874,6 +925,9 @@ async def test_chains_astream_chat_models(model_name):
874925
875926
876927# Async Streaming in completions models
928+ @pytest .mark .skip (
929+ reason = "This test suite is not properly isolated and fails flakily. TODO: Investigate why"
930+ )
877931@pytest .mark .asyncio
878932@pytest .mark .parametrize ("model_name" , ["gpt-3.5-turbo-instruct" ])
879933async def test_chains_astream_completions_models (model_name ):
0 commit comments