forked from phoenix/litellm-mirror
test batch completions on litellm proxy
This commit is contained in:
parent
31cb1be279
commit
e1f94fcbbb
1 changed files with 26 additions and 1 deletions
|
@ -4,6 +4,7 @@ import pytest
|
||||||
import asyncio
|
import asyncio
|
||||||
import aiohttp, openai
|
import aiohttp, openai
|
||||||
from openai import OpenAI, AsyncOpenAI
|
from openai import OpenAI, AsyncOpenAI
|
||||||
|
from typing import Optional, List, Union
|
||||||
|
|
||||||
|
|
||||||
def response_header_check(response):
|
def response_header_check(response):
|
||||||
|
@ -71,7 +72,7 @@ async def new_user(session):
|
||||||
return await response.json()
|
return await response.json()
|
||||||
|
|
||||||
|
|
||||||
async def chat_completion(session, key, model="gpt-4"):
|
async def chat_completion(session, key, model: Union[str, List] = "gpt-4"):
|
||||||
url = "http://0.0.0.0:4000/chat/completions"
|
url = "http://0.0.0.0:4000/chat/completions"
|
||||||
headers = {
|
headers = {
|
||||||
"Authorization": f"Bearer {key}",
|
"Authorization": f"Bearer {key}",
|
||||||
|
@ -409,3 +410,27 @@ async def test_openai_wildcard_chat_completion():
|
||||||
|
|
||||||
# call chat/completions with a model that the key was not created for + the model is not on the config.yaml
|
# call chat/completions with a model that the key was not created for + the model is not on the config.yaml
|
||||||
await chat_completion(session=session, key=key, model="gpt-3.5-turbo-0125")
|
await chat_completion(session=session, key=key, model="gpt-3.5-turbo-0125")
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.asyncio
|
||||||
|
async def test_batch_chat_completions():
|
||||||
|
"""
|
||||||
|
- Make chat completion call using
|
||||||
|
|
||||||
|
"""
|
||||||
|
async with aiohttp.ClientSession() as session:
|
||||||
|
|
||||||
|
# call chat/completions with a model that the key was not created for + the model is not on the config.yaml
|
||||||
|
response = await chat_completion(
|
||||||
|
session=session,
|
||||||
|
key="sk-1234",
|
||||||
|
model=[
|
||||||
|
"gpt-3.5-turbo",
|
||||||
|
"fake-openai-endpoint",
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
|
print(f"response: {response}")
|
||||||
|
|
||||||
|
assert len(response) == 2
|
||||||
|
assert isinstance(response, list)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue