diff --git a/litellm/tests/test_text_completion.py b/litellm/tests/test_text_completion.py index 6a093af237..20809e7c7f 100644 --- a/litellm/tests/test_text_completion.py +++ b/litellm/tests/test_text_completion.py @@ -3990,6 +3990,7 @@ def test_async_text_completion(): asyncio.run(test_get_response()) +@pytest.mark.skip(reason="Skip flaky tgai test") def test_async_text_completion_together_ai(): litellm.set_verbose = True print("test_async_text_completion") @@ -3997,7 +3998,7 @@ def test_async_text_completion_together_ai(): async def test_get_response(): try: response = await litellm.atext_completion( - model="together_ai/codellama/CodeLlama-13b-Instruct-hf", + model="together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1", prompt="good morning", max_tokens=10, ) diff --git a/litellm/utils.py b/litellm/utils.py index f2bc7b3070..52e94f28fa 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -8643,6 +8643,14 @@ def exception_type( response=original_exception.response, litellm_debug_info=extra_information, ) + elif "A timeout occurred" in error_str: + exception_mapping_worked = True + raise Timeout( + message=f"{exception_provider} - {message}", + model=model, + llm_provider=custom_llm_provider, + litellm_debug_info=extra_information, + ) elif ( "invalid_request_error" in error_str and "content_policy_violation" in error_str