test(test_custom_logger.py): reset cache test correctly

This commit is contained in:
Krrish Dholakia 2023-12-11 17:41:47 -08:00
parent 6821d74588
commit 1f18093b63
3 changed files with 15 additions and 15 deletions

1
.gitignore vendored
View file

@ -19,3 +19,4 @@ litellm/proxy/_secret_config.yaml
litellm/tests/aiologs.log litellm/tests/aiologs.log
litellm/tests/exception_data.txt litellm/tests/exception_data.txt
litellm/tests/config_*.yaml litellm/tests/config_*.yaml
litellm/tests/langfuse.log

View file

@ -258,10 +258,10 @@ class CompletionCustomHandler(CustomLogger): # https://docs.litellm.ai/docs/obse
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_async_chat_azure(): async def test_async_chat_azure():
try: try:
customHandler = CompletionCustomHandler() customHandler_completion_azure_router = CompletionCustomHandler()
customHandler_streaming = CompletionCustomHandler() customHandler_streaming_azure_router = CompletionCustomHandler()
customHandler_failure = CompletionCustomHandler() customHandler_failure = CompletionCustomHandler()
litellm.callbacks = [customHandler] litellm.callbacks = [customHandler_completion_azure_router]
model_list = [ model_list = [
{ {
"model_name": "gpt-3.5-turbo", # openai model name "model_name": "gpt-3.5-turbo", # openai model name
@ -282,10 +282,10 @@ async def test_async_chat_azure():
"content": "Hi 👋 - i'm openai" "content": "Hi 👋 - i'm openai"
}]) }])
await asyncio.sleep(2) await asyncio.sleep(2)
assert len(customHandler.errors) == 0 assert len(customHandler_completion_azure_router.errors) == 0
assert len(customHandler.states) == 3 # pre, post, success assert len(customHandler_completion_azure_router.states) == 3 # pre, post, success
# streaming # streaming
litellm.callbacks = [customHandler_streaming] litellm.callbacks = [customHandler_streaming_azure_router]
router2 = Router(model_list=model_list) # type: ignore router2 = Router(model_list=model_list) # type: ignore
response = await router2.acompletion(model="gpt-3.5-turbo", response = await router2.acompletion(model="gpt-3.5-turbo",
messages=[{ messages=[{
@ -294,11 +294,12 @@ async def test_async_chat_azure():
}], }],
stream=True) stream=True)
async for chunk in response: async for chunk in response:
print(f"async azure router chunk: {chunk}")
continue continue
await asyncio.sleep(1) await asyncio.sleep(1)
print(f"customHandler.states: {customHandler_streaming.states}") print(f"customHandler.states: {customHandler_streaming_azure_router.states}")
assert len(customHandler_streaming.errors) == 0 assert len(customHandler_streaming_azure_router.errors) == 0
assert len(customHandler_streaming.states) >= 4 # pre, post, stream (multiple times), success assert len(customHandler_streaming_azure_router.states) >= 4 # pre, post, stream (multiple times), success
# failure # failure
model_list = [ model_list = [
{ {
@ -396,6 +397,7 @@ async def test_async_embedding_azure():
async def test_async_chat_azure_with_fallbacks(): async def test_async_chat_azure_with_fallbacks():
try: try:
customHandler_fallbacks = CompletionCustomHandler() customHandler_fallbacks = CompletionCustomHandler()
litellm.callbacks = [customHandler_fallbacks]
# with fallbacks # with fallbacks
model_list = [ model_list = [
{ {
@ -428,6 +430,7 @@ async def test_async_chat_azure_with_fallbacks():
print(f"customHandler_fallbacks.states: {customHandler_fallbacks.states}") print(f"customHandler_fallbacks.states: {customHandler_fallbacks.states}")
assert len(customHandler_fallbacks.errors) == 0 assert len(customHandler_fallbacks.errors) == 0
assert len(customHandler_fallbacks.states) == 6 # pre, post, failure, pre, post, success assert len(customHandler_fallbacks.states) == 6 # pre, post, failure, pre, post, success
litellm.callbacks = []
except Exception as e: except Exception as e:
print(f"Assertion Error: {traceback.format_exc()}") print(f"Assertion Error: {traceback.format_exc()}")
pytest.fail(f"An exception occurred - {str(e)}") pytest.fail(f"An exception occurred - {str(e)}")

View file

@ -205,7 +205,6 @@ def test_azure_completion_stream():
assert response_in_success_handler == complete_streaming_response assert response_in_success_handler == complete_streaming_response
except Exception as e: except Exception as e:
pytest.fail(f"Error occurred: {e}") pytest.fail(f"Error occurred: {e}")
test_azure_completion_stream()
def test_async_custom_handler(): def test_async_custom_handler():
try: try:
@ -316,13 +315,10 @@ def test_redis_cache_completion_stream():
print("\nresponse 2", response_2_content) print("\nresponse 2", response_2_content)
assert response_1_content == response_2_content, f"Response 1 != Response 2. Same params, Response 1{response_1_content} != Response 2{response_2_content}" assert response_1_content == response_2_content, f"Response 1 != Response 2. Same params, Response 1{response_1_content} != Response 2{response_2_content}"
litellm.success_callback = [] litellm.success_callback = []
litellm._async_success_callback = []
litellm.cache = None litellm.cache = None
except Exception as e: except Exception as e:
print(e) print(e)
litellm.success_callback = [] litellm.success_callback = []
raise e raise e
""" pytest.fail(f"Error occurred: {e}")
1 & 2 should be exactly the same
"""
# test_redis_cache_completion_stream()