litellm-mirror/litellm/tests/test_model_fallback.py
2023-07-31 18:26:16 -07:00

36 lines
1.2 KiB
Python

#### What this tests ####
# This tests error handling + logging (esp. for sentry breadcrumbs)
import sys, os
import traceback
sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the system path
import litellm
from litellm import embedding, completion
litellm.success_callback = ["posthog"]
litellm.failure_callback = ["slack", "sentry", "posthog"]
litellm.set_verbose = True
model_fallback_list = ["replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1", "replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1", "chatgpt-test"]
user_message = "Hello, how are you?"
messages = [{ "content": user_message,"role": "user"}]
# for _ in range(10):
for model in model_fallback_list:
try:
response = embedding(model="text-embedding-ada-002", input=[user_message])
response = completion(model=model, messages=messages)
print(response)
<<<<<<< HEAD
if response != None:
break
except Exception as e:
=======
# if response != None:
# break
except:
>>>>>>> bd42ec8 (clean up code files)
print(f"error occurred: {traceback.format_exc()}")
raise e