(test) proxy - add openai exception mapping error

This commit is contained in:
ishaan-jaff 2024-01-15 09:56:20 -08:00
parent 746213f5d1
commit 44a7e4157c

View file

@ -0,0 +1,51 @@
import openai, httpx, os
from dotenv import load_dotenv
load_dotenv()
client = openai.OpenAI(
api_key="anything",
base_url="http://0.0.0.0:8000",
http_client=httpx.Client(verify=False),
)
try:
# request sent to model set on litellm proxy, `litellm --model`
response = client.chat.completions.create(
model="azure-gpt-3.5",
messages=[
{
"role": "user",
"content": "this is a test request, write a short poem" * 2000,
}
],
)
print(response)
except Exception as e:
print(e)
variables_proxy_exception = vars(e)
print("proxy exception variables", variables_proxy_exception.keys())
print(variables_proxy_exception["body"])
api_key = os.getenv("AZURE_API_KEY")
azure_endpoint = os.getenv("AZURE_API_BASE")
print(api_key, azure_endpoint)
client = openai.AzureOpenAI(
api_key=os.getenv("AZURE_API_KEY"),
azure_endpoint=os.getenv("AZURE_API_BASE", "default"),
)
try:
response = client.chat.completions.create(
model="chatgpt-v-2",
messages=[
{
"role": "user",
"content": "this is a test request, write a short poem" * 2000,
}
],
)
except Exception as e:
print(e)
variables_exception = vars(e)
print("openai client exception variables", variables_exception.keys())