From 1eca893d2f28d32d1cbdaa2dba851e3fdd861b50 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Tue, 24 Oct 2023 12:53:59 -0700 Subject: [PATCH] test(test_async_fn.py): adding additional logging for tests --- litellm/tests/test_async_fn.py | 2 ++ litellm/tests/test_caching.py | 1 - 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/litellm/tests/test_async_fn.py b/litellm/tests/test_async_fn.py index d0f4426b64..72e6fc3a7e 100644 --- a/litellm/tests/test_async_fn.py +++ b/litellm/tests/test_async_fn.py @@ -30,6 +30,8 @@ def test_get_response_streaming(): user_message = "Hello, how are you?" messages = [{"content": user_message, "role": "user"}] try: + import litellm + litellm.set_verbose = True response = await acompletion(model="gpt-3.5-turbo", messages=messages, stream=True) print(type(response)) diff --git a/litellm/tests/test_caching.py b/litellm/tests/test_caching.py index 706ea9a7e1..d24c6e4668 100644 --- a/litellm/tests/test_caching.py +++ b/litellm/tests/test_caching.py @@ -311,7 +311,6 @@ def test_redis_cache_completion(): def custom_get_cache_key(*args, **kwargs): # return key to use for your cache: key = kwargs.get("model", "") + str(kwargs.get("messages", "")) + str(kwargs.get("temperature", "")) + str(kwargs.get("logit_bias", "")) - print("key for cache", key) return key def test_custom_redis_cache_with_key():