From 4b636a357834e2a2fc45a7b0683344f5df918e7c Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Wed, 20 Sep 2023 08:02:59 -0700 Subject: [PATCH] new test for batch_completion_models_all_responses --- litellm/tests/test_batch_completions.py | 10 +++++++++- 1 file changed, 9 insertions(+), 1 deletion(-) diff --git a/litellm/tests/test_batch_completions.py b/litellm/tests/test_batch_completions.py index 03c5c29101..4994fa9e40 100644 --- a/litellm/tests/test_batch_completions.py +++ b/litellm/tests/test_batch_completions.py @@ -9,7 +9,7 @@ sys.path.insert( ) # Adds the parent directory to the system path from openai.error import Timeout import litellm -from litellm import batch_completion, batch_completion_models, completion +from litellm import batch_completion, batch_completion_models, completion, batch_completion_models_all_responses # litellm.set_verbose=True # def test_batch_completions(): @@ -35,6 +35,14 @@ def test_batch_completions_models(): pytest.fail(f"An error occurred: {e}") # test_batch_completions_models() +def test_batch_completion_models_all_responses(): + responses = batch_completion_models_all_responses( + models=["gpt-3.5-turbo", "claude-instant-1.2", "command-nightly"], + messages=[{"role": "user", "content": "Hey, how's it going"}], + max_tokens=5 + ) + print(responses) +# test_batch_completion_models_all_responses() # def test_batch_completions(): # try: