mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
refactor(openai.py): working openai chat + text completion for openai v1 sdk
This commit is contained in:
parent
6d815d98fe
commit
d0bd932b3c
6 changed files with 30 additions and 27 deletions
|
@ -28,14 +28,13 @@ def test_async_response():
|
|||
user_message = "Hello, how are you?"
|
||||
messages = [{"content": user_message, "role": "user"}]
|
||||
try:
|
||||
response = await acompletion(model="gpt-3.5-turbo-instruct", messages=messages)
|
||||
response = await acompletion(model="gpt-3.5-turbo", messages=messages)
|
||||
print(f"response: {response}")
|
||||
except Exception as e:
|
||||
pytest.fail(f"An exception occurred: {e}")
|
||||
|
||||
response = asyncio.run(test_get_response())
|
||||
print(response)
|
||||
# test_async_response()
|
||||
asyncio.run(test_get_response())
|
||||
test_async_response()
|
||||
|
||||
def test_get_response_streaming():
|
||||
import asyncio
|
||||
|
@ -43,7 +42,7 @@ def test_get_response_streaming():
|
|||
user_message = "write a short poem in one sentence"
|
||||
messages = [{"content": user_message, "role": "user"}]
|
||||
try:
|
||||
response = await acompletion(model="gpt-3.5-turbo-instruct", messages=messages, stream=True)
|
||||
response = await acompletion(model="gpt-3.5-turbo", messages=messages, stream=True)
|
||||
print(type(response))
|
||||
|
||||
import inspect
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue