diff --git a/litellm/tests/test_completion.py b/litellm/tests/test_completion.py index 10393a9532..1efd489e83 100644 --- a/litellm/tests/test_completion.py +++ b/litellm/tests/test_completion.py @@ -14,7 +14,7 @@ import litellm from litellm import embedding, completion, completion_cost from litellm import RateLimitError litellm.num_retries = 3 - +litellm.cache = None user_message = "Write a short poem about the sky" messages = [{"content": user_message, "role": "user"}] @@ -24,6 +24,7 @@ def logger_fn(user_model_dict): def test_completion_custom_provider_model_name(): try: + litellm.cache = None response = completion( model="together_ai/togethercomputer/llama-2-70b-chat", messages=messages, @@ -41,6 +42,7 @@ def test_completion_custom_provider_model_name(): def test_completion_claude(): litellm.set_verbose = False + litellm.cache = None litellm.AnthropicConfig(max_tokens_to_sample=200, metadata={"user_id": "1224"}) try: # test without max tokens diff --git a/litellm/tests/test_streaming.py b/litellm/tests/test_streaming.py index cd3734268d..c9b3b9a671 100644 --- a/litellm/tests/test_streaming.py +++ b/litellm/tests/test_streaming.py @@ -227,6 +227,7 @@ def streaming_format_tests(idx, chunk): def test_completion_cohere_stream_bad_key(): try: + litellm.cache = None api_key = "bad-key" messages = [ {"role": "system", "content": "You are a helpful assistant."},