From e5bc7b338a1de7d36e967673acb45fb22eb250da Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Mon, 11 Dec 2023 22:57:39 -0800 Subject: [PATCH] (test) stream - cache --- litellm/tests/test_custom_logger.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/litellm/tests/test_custom_logger.py b/litellm/tests/test_custom_logger.py index 1d373e0da..03d6410d2 100644 --- a/litellm/tests/test_custom_logger.py +++ b/litellm/tests/test_custom_logger.py @@ -110,7 +110,7 @@ def test_async_chat_openai_stream(): except Exception as e: print(e) pytest.fail(f"An error occurred - {str(e)}") -# test_async_chat_openai_stream() +test_async_chat_openai_stream() def test_completion_azure_stream_moderation_failure(): try: @@ -317,11 +317,12 @@ def test_redis_cache_completion_stream(): response_2_content += chunk.choices[0].delta.content or "" print("\nresponse 1", response_1_content) print("\nresponse 2", response_2_content) - assert response1.id == response2.id, f"Response 1 != Response 2. Same params, Response 1{response1.id} != Response 2{response2.id}" + assert response_1_content == response_2_content, f"Response 1 != Response 2. Same params, Response 1{response_1_content} != Response 2{response_2_content}" litellm.success_callback = [] litellm._async_success_callback = [] litellm.cache = None except Exception as e: print(e) litellm.success_callback = [] - raise e \ No newline at end of file + raise e +test_redis_cache_completion_stream() \ No newline at end of file