test(test_router_fallbacks.py): update to use async logging

This commit is contained in:
Krrish Dholakia 2023-12-11 20:12:21 -08:00
parent da0ec40a9f
commit b3600e9e8a
3 changed files with 56 additions and 39 deletions

View file

@ -1,5 +1,5 @@
Task exception was never retrieved Task exception was never retrieved
future: <Task finished name='Task-295' coro=<QueryEngine.aclose() done, defined at /opt/homebrew/lib/python3.11/site-packages/prisma/engine/query.py:110> exception=RuntimeError('Event loop is closed')> future: <Task finished name='Task-299' coro=<QueryEngine.aclose() done, defined at /opt/homebrew/lib/python3.11/site-packages/prisma/engine/query.py:110> exception=RuntimeError('Event loop is closed')>
Traceback (most recent call last): Traceback (most recent call last):
File "/opt/homebrew/lib/python3.11/site-packages/prisma/engine/query.py", line 112, in aclose File "/opt/homebrew/lib/python3.11/site-packages/prisma/engine/query.py", line 112, in aclose
await self._close_session() await self._close_session()
@ -29,3 +29,5 @@ Traceback (most recent call last):
raise RuntimeError('Event loop is closed') raise RuntimeError('Event loop is closed')
RuntimeError: Event loop is closed RuntimeError: Event loop is closed
Giving up get_data(...) after 3 tries (prisma.errors.ClientNotConnectedError: Client is not connected to the query engine, you must call `connect()` before attempting to query data.) Giving up get_data(...) after 3 tries (prisma.errors.ClientNotConnectedError: Client is not connected to the query engine, you must call `connect()` before attempting to query data.)
Giving up get_data(...) after 3 tries (prisma.errors.ClientNotConnectedError: Client is not connected to the query engine, you must call `connect()` before attempting to query data.)
Giving up get_data(...) after 3 tries (prisma.errors.ClientNotConnectedError: Client is not connected to the query engine, you must call `connect()` before attempting to query data.)

View file

@ -43,3 +43,15 @@ model_list:
description: this is a test openai model description: this is a test openai model
id: 79fc75bf-8e1b-47d5-8d24-9365a854af03 id: 79fc75bf-8e1b-47d5-8d24-9365a854af03
model_name: test_openai_models model_name: test_openai_models
- litellm_params:
model: gpt-3.5-turbo
model_info:
description: this is a test openai model
id: 78ec9388-427e-49c7-bdf2-355f00a6b246
model_name: test_openai_models
- litellm_params:
model: gpt-3.5-turbo
model_info:
description: this is a test openai model
id: cb373690-360e-4aba-8fae-71de12f5278e
model_name: test_openai_models

View file

@ -31,6 +31,12 @@ class MyCustomHandler(CustomLogger):
self.previous_models += len(kwargs["litellm_params"]["metadata"]["previous_models"]) # {"previous_models": [{"model": litellm_model_name, "exception_type": AuthenticationError, "exception_string": <complete_traceback>}]} self.previous_models += len(kwargs["litellm_params"]["metadata"]["previous_models"]) # {"previous_models": [{"model": litellm_model_name, "exception_type": AuthenticationError, "exception_string": <complete_traceback>}]}
print(f"self.previous_models: {self.previous_models}") print(f"self.previous_models: {self.previous_models}")
print(f"On Success") print(f"On Success")
async def async_log_success_event(self, kwargs, response_obj, start_time, end_time):
print(f"previous_models: {kwargs['litellm_params']['metadata']['previous_models']}")
self.previous_models += len(kwargs["litellm_params"]["metadata"]["previous_models"]) # {"previous_models": [{"model": litellm_model_name, "exception_type": AuthenticationError, "exception_string": <complete_traceback>}]}
print(f"self.previous_models: {self.previous_models}")
print(f"On Success")
def log_failure_event(self, kwargs, response_obj, start_time, end_time): def log_failure_event(self, kwargs, response_obj, start_time, end_time):
print(f"On Failure") print(f"On Failure")
@ -111,30 +117,29 @@ def test_sync_fallbacks():
print(e) print(e)
# test_sync_fallbacks() # test_sync_fallbacks()
def test_async_fallbacks(): @pytest.mark.asyncio
async def test_async_fallbacks():
litellm.set_verbose = False litellm.set_verbose = False
router = Router(model_list=model_list, router = Router(model_list=model_list,
fallbacks=[{"azure/gpt-3.5-turbo": ["gpt-3.5-turbo"]}], fallbacks=[{"azure/gpt-3.5-turbo": ["gpt-3.5-turbo"]}],
context_window_fallbacks=[{"azure/gpt-3.5-turbo-context-fallback": ["gpt-3.5-turbo-16k"]}, {"gpt-3.5-turbo": ["gpt-3.5-turbo-16k"]}], context_window_fallbacks=[{"azure/gpt-3.5-turbo-context-fallback": ["gpt-3.5-turbo-16k"]}, {"gpt-3.5-turbo": ["gpt-3.5-turbo-16k"]}],
set_verbose=False) set_verbose=False)
async def test_get_response(): customHandler = MyCustomHandler()
customHandler = MyCustomHandler() litellm.callbacks = [customHandler]
litellm.callbacks = [customHandler] user_message = "Hello, how are you?"
user_message = "Hello, how are you?" messages = [{"content": user_message, "role": "user"}]
messages = [{"content": user_message, "role": "user"}] try:
try: response = await router.acompletion(**kwargs)
response = await router.acompletion(**kwargs) print(f"customHandler.previous_models: {customHandler.previous_models}")
print(f"customHandler.previous_models: {customHandler.previous_models}") await asyncio.sleep(0.05) # allow a delay as success_callbacks are on a separate thread
time.sleep(0.05) # allow a delay as success_callbacks are on a separate thread assert customHandler.previous_models == 1 # 0 retries, 1 fallback
assert customHandler.previous_models == 1 # 0 retries, 1 fallback router.reset()
router.reset() except litellm.Timeout as e:
except litellm.Timeout as e: pass
pass except Exception as e:
except Exception as e: pytest.fail(f"An exception occurred: {e}")
pytest.fail(f"An exception occurred: {e}") finally:
finally: router.reset()
router.reset()
asyncio.run(test_get_response())
# test_async_fallbacks() # test_async_fallbacks()
@ -184,26 +189,24 @@ def test_dynamic_fallbacks_sync():
# test_dynamic_fallbacks_sync() # test_dynamic_fallbacks_sync()
def test_dynamic_fallbacks_async(): @pytest.mark.asyncio
async def test_dynamic_fallbacks_async():
""" """
Allow setting the fallback in the router.completion() call. Allow setting the fallback in the router.completion() call.
""" """
async def test_get_response(): try:
try: customHandler = MyCustomHandler()
customHandler = MyCustomHandler() litellm.callbacks = [customHandler]
litellm.callbacks = [customHandler] router = Router(model_list=model_list, set_verbose=True)
router = Router(model_list=model_list, set_verbose=True) kwargs = {}
kwargs = {} kwargs["model"] = "azure/gpt-3.5-turbo"
kwargs["model"] = "azure/gpt-3.5-turbo" kwargs["messages"] = [{"role": "user", "content": "Hey, how's it going?"}]
kwargs["messages"] = [{"role": "user", "content": "Hey, how's it going?"}] kwargs["fallbacks"] = [{"azure/gpt-3.5-turbo": ["gpt-3.5-turbo"]}]
kwargs["fallbacks"] = [{"azure/gpt-3.5-turbo": ["gpt-3.5-turbo"]}] response = await router.acompletion(**kwargs)
response = await router.acompletion(**kwargs) print(f"response: {response}")
print(f"response: {response}") await asyncio.sleep(0.05) # allow a delay as success_callbacks are on a separate thread
time.sleep(0.05) # allow a delay as success_callbacks are on a separate thread assert customHandler.previous_models == 1 # 0 retries, 1 fallback
assert customHandler.previous_models == 1 # 0 retries, 1 fallback router.reset()
router.reset() except Exception as e:
except Exception as e: pytest.fail(f"An exception occurred - {e}")
pytest.fail(f"An exception occurred - {e}")
asyncio.run(test_get_response())
# test_dynamic_fallbacks_async() # test_dynamic_fallbacks_async()