From 914cec3ab5e0c931172c812a6d8a9214a5f49bba Mon Sep 17 00:00:00 2001 From: Ishaan Jaff Date: Wed, 13 Nov 2024 19:37:34 -0800 Subject: [PATCH] test - handle eol model claude-2, use claude-2.1 instead --- tests/local_testing/test_router_fallbacks.py | 14 +++++++------- tests/local_testing/test_streaming.py | 2 +- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/tests/local_testing/test_router_fallbacks.py b/tests/local_testing/test_router_fallbacks.py index cad640a54..7f4d318bf 100644 --- a/tests/local_testing/test_router_fallbacks.py +++ b/tests/local_testing/test_router_fallbacks.py @@ -1138,9 +1138,9 @@ async def test_router_content_policy_fallbacks( router = Router( model_list=[ { - "model_name": "claude-2", + "model_name": "claude-2.1", "litellm_params": { - "model": "claude-2", + "model": "claude-2.1", "api_key": "", "mock_response": mock_response, }, @@ -1164,7 +1164,7 @@ async def test_router_content_policy_fallbacks( { "model_name": "my-general-model", "litellm_params": { - "model": "claude-2", + "model": "claude-2.1", "api_key": "", "mock_response": Exception("Should not have called this."), }, @@ -1172,14 +1172,14 @@ async def test_router_content_policy_fallbacks( { "model_name": "my-context-window-model", "litellm_params": { - "model": "claude-2", + "model": "claude-2.1", "api_key": "", "mock_response": Exception("Should not have called this."), }, }, ], content_policy_fallbacks=( - [{"claude-2": ["my-fallback-model"]}] + [{"claude-2.1": ["my-fallback-model"]}] if fallback_type == "model-specific" else None ), @@ -1190,12 +1190,12 @@ async def test_router_content_policy_fallbacks( if sync_mode is True: response = router.completion( - model="claude-2", + model="claude-2.1", messages=[{"role": "user", "content": "Hey, how's it going?"}], ) else: response = await router.acompletion( - model="claude-2", + model="claude-2.1", messages=[{"role": "user", "content": "Hey, how's it going?"}], ) diff --git a/tests/local_testing/test_streaming.py b/tests/local_testing/test_streaming.py index fcdc6b60d..7b3dbd8d6 100644 --- a/tests/local_testing/test_streaming.py +++ b/tests/local_testing/test_streaming.py @@ -718,7 +718,7 @@ async def test_acompletion_claude_2_stream(): try: litellm.set_verbose = True response = await litellm.acompletion( - model="claude-2", + model="claude-2.1", messages=[{"role": "user", "content": "hello from litellm"}], stream=True, )