forked from phoenix/litellm-mirror
cleanup
This commit is contained in:
parent
d77d8b7cd7
commit
c050258530
12 changed files with 17 additions and 8 deletions
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
|
@ -26,7 +26,7 @@ litellm.failure_callback = ["slack", "sentry", "posthog"]
|
||||||
|
|
||||||
user_message = "Hello, how are you?"
|
user_message = "Hello, how are you?"
|
||||||
messages = [{ "content": user_message,"role": "user"}]
|
messages = [{ "content": user_message,"role": "user"}]
|
||||||
model_val = "krrish is a model"
|
model_val = None
|
||||||
|
|
||||||
|
|
||||||
def test_completion_with_empty_model():
|
def test_completion_with_empty_model():
|
||||||
|
@ -35,4 +35,4 @@ def test_completion_with_empty_model():
|
||||||
response = completion(model=model_val, messages=messages)
|
response = completion(model=model_val, messages=messages)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"error occurred: {e}")
|
print(f"error occurred: {e}")
|
||||||
pass
|
pass
|
|
@ -1,5 +1,8 @@
|
||||||
import sys, os
|
import sys, os
|
||||||
import traceback
|
import traceback
|
||||||
|
from dotenv import load_dotenv
|
||||||
|
load_dotenv()
|
||||||
|
import os
|
||||||
sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the system path
|
sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the system path
|
||||||
import pytest
|
import pytest
|
||||||
import litellm
|
import litellm
|
||||||
|
|
|
@ -7,7 +7,9 @@ sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the
|
||||||
import litellm
|
import litellm
|
||||||
from litellm import embedding, completion
|
from litellm import embedding, completion
|
||||||
|
|
||||||
litellm.set_verbose = True
|
litellm.set_verbose = False
|
||||||
|
|
||||||
|
score = 0
|
||||||
|
|
||||||
def logger_fn(model_call_object: dict):
|
def logger_fn(model_call_object: dict):
|
||||||
print(f"model call details: {model_call_object}")
|
print(f"model call details: {model_call_object}")
|
||||||
|
@ -18,6 +20,7 @@ messages = [{ "content": user_message,"role": "user"}]
|
||||||
# test on openai completion call
|
# test on openai completion call
|
||||||
try:
|
try:
|
||||||
response = completion(model="gpt-3.5-turbo", messages=messages)
|
response = completion(model="gpt-3.5-turbo", messages=messages)
|
||||||
|
score +=1
|
||||||
except:
|
except:
|
||||||
print(f"error occurred: {traceback.format_exc()}")
|
print(f"error occurred: {traceback.format_exc()}")
|
||||||
pass
|
pass
|
||||||
|
@ -25,6 +28,7 @@ except:
|
||||||
# test on non-openai completion call
|
# test on non-openai completion call
|
||||||
try:
|
try:
|
||||||
response = completion(model="claude-instant-1", messages=messages, logger_fn=logger_fn)
|
response = completion(model="claude-instant-1", messages=messages, logger_fn=logger_fn)
|
||||||
|
score +=1
|
||||||
except:
|
except:
|
||||||
print(f"error occurred: {traceback.format_exc()}")
|
print(f"error occurred: {traceback.format_exc()}")
|
||||||
pass
|
pass
|
||||||
|
@ -32,20 +36,23 @@ except:
|
||||||
# test on openai embedding call
|
# test on openai embedding call
|
||||||
try:
|
try:
|
||||||
response = embedding(model='text-embedding-ada-002', input=[user_message], logger_fn=logger_fn)
|
response = embedding(model='text-embedding-ada-002', input=[user_message], logger_fn=logger_fn)
|
||||||
print(f"response: {str(response)[:50]}")
|
score +=1
|
||||||
except:
|
except:
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
|
|
||||||
# test on bad azure openai embedding call -> missing azure flag and this isn't an embedding model
|
# test on bad azure openai embedding call -> missing azure flag and this isn't an embedding model
|
||||||
try:
|
try:
|
||||||
response = embedding(model='chatgpt-test', input=[user_message], logger_fn=logger_fn)
|
response = embedding(model='chatgpt-test', input=[user_message], logger_fn=logger_fn)
|
||||||
print(f"response: {str(response)[:50]}")
|
|
||||||
except:
|
except:
|
||||||
|
score +=1 # expect this to fail
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
|
|
||||||
# test on good azure openai embedding call
|
# test on good azure openai embedding call
|
||||||
try:
|
try:
|
||||||
response = embedding(model='azure-embedding-model', input=[user_message], azure=True, logger_fn=logger_fn)
|
response = embedding(model='azure-embedding-model', input=[user_message], azure=True, logger_fn=logger_fn)
|
||||||
print(f"response: {str(response)[:50]}")
|
score +=1
|
||||||
except:
|
except:
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
|
|
||||||
|
|
||||||
|
print(f"Score: {score}, Overall score: {score/5}")
|
|
@ -12,7 +12,7 @@ litellm.failure_callback = ["slack", "sentry", "posthog"]
|
||||||
|
|
||||||
litellm.set_verbose = True
|
litellm.set_verbose = True
|
||||||
|
|
||||||
model_fallback_list = ["replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1", "replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1", "chatgpt-test"]
|
model_fallback_list = ["claude-instant-1", "gpt-3.5-turbo", "chatgpt-test"]
|
||||||
|
|
||||||
user_message = "Hello, how are you?"
|
user_message = "Hello, how are you?"
|
||||||
messages = [{ "content": user_message,"role": "user"}]
|
messages = [{ "content": user_message,"role": "user"}]
|
||||||
|
@ -21,6 +21,5 @@ for model in model_fallback_list:
|
||||||
try:
|
try:
|
||||||
response = embedding(model="text-embedding-ada-002", input=[user_message])
|
response = embedding(model="text-embedding-ada-002", input=[user_message])
|
||||||
response = completion(model=model, messages=messages)
|
response = completion(model=model, messages=messages)
|
||||||
print(response)
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"error occurred: {traceback.format_exc()}")
|
print(f"error occurred: {traceback.format_exc()}")
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue