From 9bde3ccd1d3cba67462ff472b105fc53d357d131 Mon Sep 17 00:00:00 2001 From: Ishaan Jaff Date: Mon, 13 May 2024 20:48:38 -0700 Subject: [PATCH] (ci/cd) fixes --- litellm/tests/test_completion.py | 2 +- proxy_server_config.yaml | 1 - 2 files changed, 1 insertion(+), 2 deletions(-) diff --git a/litellm/tests/test_completion.py b/litellm/tests/test_completion.py index aa0baa2b4..ccc73a260 100644 --- a/litellm/tests/test_completion.py +++ b/litellm/tests/test_completion.py @@ -885,7 +885,7 @@ async def test_acompletion_claude2_1(): }, {"role": "user", "content": "Generate a 3 liner joke for me"}, ] - # test without max tokens + # test without max-tokens response = await litellm.acompletion(model="claude-2.1", messages=messages) # Add any assertions here to check the response print(response) diff --git a/proxy_server_config.yaml b/proxy_server_config.yaml index 0673d967d..9b0e7c9d0 100644 --- a/proxy_server_config.yaml +++ b/proxy_server_config.yaml @@ -84,7 +84,6 @@ model_list: model: text-completion-openai/gpt-3.5-turbo-instruct litellm_settings: drop_params: True - enable_preview_features: True # max_budget: 100 # budget_duration: 30d num_retries: 5