fix(proxy_server.py): accept max tokens as int

This commit is contained in:
Krrish Dholakia 2023-12-01 15:19:34 -08:00
parent ac4429c35f
commit 4da20ac0b9
2 changed files with 3 additions and 3 deletions

View file

@ -126,7 +126,7 @@ class ProxyChatCompletionRequest(BaseModel):
n: Optional[int] = None n: Optional[int] = None
stream: Optional[bool] = None stream: Optional[bool] = None
stop: Optional[List[str]] = None stop: Optional[List[str]] = None
max_tokens: Optional[float] = None max_tokens: Optional[Union[float, int]] = None
presence_penalty: Optional[float] = None presence_penalty: Optional[float] = None
frequency_penalty: Optional[float] = None frequency_penalty: Optional[float] = None
logit_bias: Optional[Dict[str, float]] = None logit_bias: Optional[Dict[str, float]] = None
@ -752,7 +752,7 @@ async def completion(request: Request, model: Optional[str] = None, user_api_key
@router.post("/v1/chat/completions", dependencies=[Depends(user_api_key_auth)], tags=["chat/completions"]) @router.post("/v1/chat/completions", dependencies=[Depends(user_api_key_auth)], tags=["chat/completions"])
@router.post("/chat/completions", dependencies=[Depends(user_api_key_auth)], tags=["chat/completions"]) @router.post("/chat/completions", dependencies=[Depends(user_api_key_auth)], tags=["chat/completions"])
@router.post("/openai/deployments/{model:path}/chat/completions", dependencies=[Depends(user_api_key_auth)], tags=["chat/completions"]) # azure compatible endpoint @router.post("/openai/deployments/{model:path}/chat/completions", dependencies=[Depends(user_api_key_auth)], tags=["chat/completions"]) # azure compatible endpoint
async def chat_completion(request: ProxyChatCompletionRequest, model: Optional[str] = None, user_api_key_dict: dict = Depends(user_api_key_auth)) -> Union[litellm.ModelResponse, StreamingResponse]: async def chat_completion(request: ProxyChatCompletionRequest, model: Optional[str] = None, user_api_key_dict: dict = Depends(user_api_key_auth)):
global general_settings, user_debug global general_settings, user_debug
try: try:
data = {} data = {}

View file

@ -45,7 +45,7 @@ def test_chat_completion():
pytest.fail("LiteLLM Proxy test failed. Exception", e) pytest.fail("LiteLLM Proxy test failed. Exception", e)
# Run the test # Run the test
# test_chat_completion() test_chat_completion()
def test_chat_completion_azure(): def test_chat_completion_azure():