diff --git a/dist/litellm-1.0.3.dev1-py3-none-any.whl b/dist/litellm-1.0.3.dev1-py3-none-any.whl new file mode 100644 index 000000000..8a316ed32 Binary files /dev/null and b/dist/litellm-1.0.3.dev1-py3-none-any.whl differ diff --git a/dist/litellm-1.0.3.dev1.tar.gz b/dist/litellm-1.0.3.dev1.tar.gz new file mode 100644 index 000000000..8ff781620 Binary files /dev/null and b/dist/litellm-1.0.3.dev1.tar.gz differ diff --git a/litellm/tests/test_completion.py b/litellm/tests/test_completion.py index ed1e49b3b..a7af44534 100644 --- a/litellm/tests/test_completion.py +++ b/litellm/tests/test_completion.py @@ -1410,7 +1410,7 @@ def test_completion_deep_infra_mistral(): def test_completion_palm(): litellm.set_verbose = True model_name = "palm/chat-bison" - messages = [{"role": "user", "content": "what model are you?"}] + messages = [{"role": "user", "content": "Hey, how's it going?"}] try: response = completion(model=model_name, messages=messages) # Add any assertions here to check the response