mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
(test) proxy - add openai exception mapping error
This commit is contained in:
parent
746213f5d1
commit
44a7e4157c
1 changed files with 51 additions and 0 deletions
51
litellm/proxy/tests/test_openai_exception_request.py
Normal file
51
litellm/proxy/tests/test_openai_exception_request.py
Normal file
|
@ -0,0 +1,51 @@
|
|||
import openai, httpx, os
|
||||
from dotenv import load_dotenv
|
||||
|
||||
load_dotenv()
|
||||
client = openai.OpenAI(
|
||||
api_key="anything",
|
||||
base_url="http://0.0.0.0:8000",
|
||||
http_client=httpx.Client(verify=False),
|
||||
)
|
||||
|
||||
try:
|
||||
# request sent to model set on litellm proxy, `litellm --model`
|
||||
response = client.chat.completions.create(
|
||||
model="azure-gpt-3.5",
|
||||
messages=[
|
||||
{
|
||||
"role": "user",
|
||||
"content": "this is a test request, write a short poem" * 2000,
|
||||
}
|
||||
],
|
||||
)
|
||||
|
||||
print(response)
|
||||
except Exception as e:
|
||||
print(e)
|
||||
variables_proxy_exception = vars(e)
|
||||
print("proxy exception variables", variables_proxy_exception.keys())
|
||||
print(variables_proxy_exception["body"])
|
||||
|
||||
|
||||
api_key = os.getenv("AZURE_API_KEY")
|
||||
azure_endpoint = os.getenv("AZURE_API_BASE")
|
||||
print(api_key, azure_endpoint)
|
||||
client = openai.AzureOpenAI(
|
||||
api_key=os.getenv("AZURE_API_KEY"),
|
||||
azure_endpoint=os.getenv("AZURE_API_BASE", "default"),
|
||||
)
|
||||
try:
|
||||
response = client.chat.completions.create(
|
||||
model="chatgpt-v-2",
|
||||
messages=[
|
||||
{
|
||||
"role": "user",
|
||||
"content": "this is a test request, write a short poem" * 2000,
|
||||
}
|
||||
],
|
||||
)
|
||||
except Exception as e:
|
||||
print(e)
|
||||
variables_exception = vars(e)
|
||||
print("openai client exception variables", variables_exception.keys())
|
Loading…
Add table
Add a link
Reference in a new issue