mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
refactor(openai.py): moving openai chat completion calls to http
This commit is contained in:
parent
da1451e493
commit
c57ed0a9d7
6 changed files with 158 additions and 127 deletions
|
@ -95,20 +95,6 @@ def test_completion_claude():
|
|||
# pytest.fail(f"Error occurred: {e}")
|
||||
# test_completion_aleph_alpha_control_models()
|
||||
|
||||
def test_completion_with_litellm_call_id():
|
||||
try:
|
||||
litellm.use_client = False
|
||||
response = completion(
|
||||
model="gpt-3.5-turbo", messages=messages)
|
||||
print(response)
|
||||
if 'litellm_call_id' in response:
|
||||
pytest.fail(f"Error occurred: litellm_call_id in response objects")
|
||||
print(response.usage)
|
||||
print(response.usage.completion_tokens)
|
||||
|
||||
except Exception as e:
|
||||
pytest.fail(f"Error occurred: {e}")
|
||||
# test_completion_with_litellm_call_id()
|
||||
import openai
|
||||
def test_completion_gpt4_turbo():
|
||||
try:
|
||||
|
@ -381,6 +367,8 @@ def test_completion_cohere(): # commenting for now as the cohere endpoint is bei
|
|||
|
||||
def test_completion_openai():
|
||||
try:
|
||||
litellm.set_verbose=True
|
||||
print(f"api key: {os.environ['OPENAI_API_KEY']}")
|
||||
litellm.api_key = os.environ['OPENAI_API_KEY']
|
||||
response = completion(model="gpt-3.5-turbo", messages=messages, max_tokens=10, request_timeout=10)
|
||||
print("This is the response object\n", response)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue