forked from phoenix/litellm-mirror
34 lines
830 B
Python
34 lines
830 B
Python
import os
|
|
from openai import OpenAI
|
|
from dotenv import load_dotenv
|
|
import httpx
|
|
import concurrent.futures
|
|
|
|
load_dotenv()
|
|
|
|
client = OpenAI(
|
|
# This is the default and can be omitted
|
|
api_key=os.environ.get("OPENAI_API_KEY"),
|
|
)
|
|
|
|
|
|
def create_chat_completion():
|
|
return client.chat.completions.create(
|
|
messages=[
|
|
{
|
|
"role": "user",
|
|
"content": "Say this is a test. Respond in 20 lines",
|
|
}
|
|
],
|
|
model="gpt-3.5-turbo",
|
|
)
|
|
|
|
|
|
with concurrent.futures.ThreadPoolExecutor() as executor:
|
|
# Set a timeout of 10 seconds
|
|
future = executor.submit(create_chat_completion)
|
|
try:
|
|
chat_completion = future.result(timeout=0.00001)
|
|
print(chat_completion)
|
|
except concurrent.futures.TimeoutError:
|
|
print("Operation timed out.")
|