diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 17fd286fd2..cf55ed638c 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -1954,7 +1954,7 @@ "litellm_provider": "mistral", "mode": "embedding" }, - "deepseek-chat": { + "deepseek/deepseek-chat": { "max_tokens": 4096, "max_input_tokens": 128000, "max_output_tokens": 4096, @@ -2023,7 +2023,7 @@ "supports_function_calling": true, "supports_vision": true }, - "deepseek-coder": { + "deepseek/deepseek-coder": { "max_tokens": 4096, "max_input_tokens": 128000, "max_output_tokens": 4096, diff --git a/tests/llm_translation/test_aiohttp_openai.py b/tests/llm_translation/test_aiohttp_openai.py index 16155bacf5..68b667c99d 100644 --- a/tests/llm_translation/test_aiohttp_openai.py +++ b/tests/llm_translation/test_aiohttp_openai.py @@ -21,4 +21,3 @@ async def test_aiohttp_openai(): api_key="fake-key", ) print(response) - print(response.model_dump_json(indent=4))