mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
(test) raise NotFoundError
This commit is contained in:
parent
0530d16595
commit
1af258ac2f
1 changed files with 13 additions and 11 deletions
|
@ -21,6 +21,7 @@ from concurrent.futures import ThreadPoolExecutor
|
||||||
import pytest
|
import pytest
|
||||||
litellm.vertex_project = "pathrise-convert-1606954137718"
|
litellm.vertex_project = "pathrise-convert-1606954137718"
|
||||||
litellm.vertex_location = "us-central1"
|
litellm.vertex_location = "us-central1"
|
||||||
|
litellm.num_retries=0
|
||||||
|
|
||||||
# litellm.failure_callback = ["sentry"]
|
# litellm.failure_callback = ["sentry"]
|
||||||
#### What this tests ####
|
#### What this tests ####
|
||||||
|
@ -38,10 +39,11 @@ models = ["command-nightly"]
|
||||||
# Test 1: Context Window Errors
|
# Test 1: Context Window Errors
|
||||||
@pytest.mark.parametrize("model", models)
|
@pytest.mark.parametrize("model", models)
|
||||||
def test_context_window(model):
|
def test_context_window(model):
|
||||||
|
print("Testing context window error")
|
||||||
sample_text = "Say error 50 times" * 1000000
|
sample_text = "Say error 50 times" * 1000000
|
||||||
messages = [{"content": sample_text, "role": "user"}]
|
messages = [{"content": sample_text, "role": "user"}]
|
||||||
try:
|
try:
|
||||||
litellm.set_verbose = False
|
litellm.set_verbose = True
|
||||||
response = completion(model=model, messages=messages)
|
response = completion(model=model, messages=messages)
|
||||||
print(f"response: {response}")
|
print(f"response: {response}")
|
||||||
print("FAILED!")
|
print("FAILED!")
|
||||||
|
@ -176,7 +178,7 @@ def test_completion_azure_exception():
|
||||||
try:
|
try:
|
||||||
import openai
|
import openai
|
||||||
print("azure gpt-3.5 test\n\n")
|
print("azure gpt-3.5 test\n\n")
|
||||||
litellm.set_verbose=False
|
litellm.set_verbose=True
|
||||||
## Test azure call
|
## Test azure call
|
||||||
old_azure_key = os.environ["AZURE_API_KEY"]
|
old_azure_key = os.environ["AZURE_API_KEY"]
|
||||||
os.environ["AZURE_API_KEY"] = "good morning"
|
os.environ["AZURE_API_KEY"] = "good morning"
|
||||||
|
@ -204,7 +206,7 @@ async def asynctest_completion_azure_exception():
|
||||||
import openai
|
import openai
|
||||||
import litellm
|
import litellm
|
||||||
print("azure gpt-3.5 test\n\n")
|
print("azure gpt-3.5 test\n\n")
|
||||||
litellm.set_verbose=False
|
litellm.set_verbose=True
|
||||||
## Test azure call
|
## Test azure call
|
||||||
old_azure_key = os.environ["AZURE_API_KEY"]
|
old_azure_key = os.environ["AZURE_API_KEY"]
|
||||||
os.environ["AZURE_API_KEY"] = "good morning"
|
os.environ["AZURE_API_KEY"] = "good morning"
|
||||||
|
@ -251,8 +253,8 @@ def asynctest_completion_openai_exception_bad_model():
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
asyncio.run(test())
|
asyncio.run(test())
|
||||||
except openai.BadRequestError:
|
except openai.NotFoundError:
|
||||||
print("Good job this is a bad request error for a model that does not exist!")
|
print("Good job this is a NotFoundError for a model that does not exist!")
|
||||||
print("Passed")
|
print("Passed")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print("Raised wrong type of exception", type(e))
|
print("Raised wrong type of exception", type(e))
|
||||||
|
@ -281,8 +283,8 @@ def asynctest_completion_azure_exception_bad_model():
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
asyncio.run(test())
|
asyncio.run(test())
|
||||||
except openai.BadRequestError:
|
except openai.NotFoundError:
|
||||||
print("Good job this is a bad request error for a model that does not exist!")
|
print("Good job this is a NotFoundError for a model that does not exist!")
|
||||||
print("Passed")
|
print("Passed")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print("Raised wrong type of exception", type(e))
|
print("Raised wrong type of exception", type(e))
|
||||||
|
@ -295,7 +297,7 @@ def test_completion_openai_exception():
|
||||||
try:
|
try:
|
||||||
import openai
|
import openai
|
||||||
print("openai gpt-3.5 test\n\n")
|
print("openai gpt-3.5 test\n\n")
|
||||||
litellm.set_verbose=False
|
litellm.set_verbose=True
|
||||||
## Test azure call
|
## Test azure call
|
||||||
old_azure_key = os.environ["OPENAI_API_KEY"]
|
old_azure_key = os.environ["OPENAI_API_KEY"]
|
||||||
os.environ["OPENAI_API_KEY"] = "good morning"
|
os.environ["OPENAI_API_KEY"] = "good morning"
|
||||||
|
@ -312,7 +314,7 @@ def test_completion_openai_exception():
|
||||||
print(response)
|
print(response)
|
||||||
except openai.AuthenticationError as e:
|
except openai.AuthenticationError as e:
|
||||||
os.environ["OPENAI_API_KEY"] = old_azure_key
|
os.environ["OPENAI_API_KEY"] = old_azure_key
|
||||||
print("good job got the correct error for openai when key not set")
|
print("OpenAI: good job got the correct error for openai when key not set")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
# test_completion_openai_exception()
|
# test_completion_openai_exception()
|
||||||
|
@ -322,7 +324,7 @@ def test_completion_mistral_exception():
|
||||||
try:
|
try:
|
||||||
import openai
|
import openai
|
||||||
print("Testing mistral ai exception mapping")
|
print("Testing mistral ai exception mapping")
|
||||||
litellm.set_verbose=False
|
litellm.set_verbose=True
|
||||||
## Test azure call
|
## Test azure call
|
||||||
old_azure_key = os.environ["MISTRAL_API_KEY"]
|
old_azure_key = os.environ["MISTRAL_API_KEY"]
|
||||||
os.environ["MISTRAL_API_KEY"] = "good morning"
|
os.environ["MISTRAL_API_KEY"] = "good morning"
|
||||||
|
@ -385,4 +387,4 @@ def test_completion_mistral_exception():
|
||||||
# counts[result] += 1
|
# counts[result] += 1
|
||||||
|
|
||||||
# accuracy_score = counts[True]/(counts[True] + counts[False])
|
# accuracy_score = counts[True]/(counts[True] + counts[False])
|
||||||
# print(f"accuracy_score: {accuracy_score}")
|
# print(f"accuracy_score: {accuracy_score}")
|
Loading…
Add table
Add a link
Reference in a new issue