From 07761ac93fab27e1a98a194c12ab70b572a0743f Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Mon, 6 Nov 2023 13:17:49 -0800 Subject: [PATCH] (test) text_completion --- litellm/tests/test_text_completion.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/litellm/tests/test_text_completion.py b/litellm/tests/test_text_completion.py index 75fd3b732a..e674a6156b 100644 --- a/litellm/tests/test_text_completion.py +++ b/litellm/tests/test_text_completion.py @@ -72,7 +72,8 @@ def test_text_completion_basic(): # print("\n testing hf mistral\n") # response = text_completion( # model="huggingface/mistralai/Mistral-7B-v0.1", -# prompt=token_prompt, # token prompt is a 2d list +# prompt=token_prompt, # token prompt is a 2d list, +# max_tokens=0.0 # ) # print("\n\n response") @@ -98,7 +99,7 @@ def test_completion_text_003_prompt_array(): # response_str = response["choices"][0]["text"] except Exception as e: pytest.fail(f"Error occurred: {e}") -test_completion_text_003_prompt_array() +# test_completion_text_003_prompt_array() # not including this in our ci cd pipeline, since we don't want to fail tests due to an unstable replit