test proxy all model

This commit is contained in:
Ishaan Jaff 2024-07-25 18:54:30 -07:00
parent 9863520376
commit 05858cb249

View file

@ -7,6 +7,9 @@ from openai import OpenAI, AsyncOpenAI
from typing import Optional, List, Union
LITELLM_MASTER_KEY = "sk-1234"
def response_header_check(response):
"""
- assert if response headers < 4kb (nginx limit).
@ -467,6 +470,22 @@ async def test_openai_wildcard_chat_completion():
await chat_completion(session=session, key=key, model="gpt-3.5-turbo-0125")
@pytest.mark.asyncio
async def test_proxy_all_models():
"""
- proxy_server_config.yaml has model = * / *
- Make chat completion call
- groq is NOT defined on /models
"""
async with aiohttp.ClientSession() as session:
# call chat/completions with a model that the key was not created for + the model is not on the config.yaml
await chat_completion(
session=session, key=LITELLM_MASTER_KEY, model="groq/llama3-8b-8192"
)
@pytest.mark.asyncio
async def test_batch_chat_completions():
"""