mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix(proxy_server.py): fix linting issues
This commit is contained in:
parent
740460f390
commit
1bef2c62a6
1 changed files with 5 additions and 5 deletions
|
@ -509,8 +509,8 @@ def model_list():
|
||||||
@router.post("/engines/{model:path}/completions")
|
@router.post("/engines/{model:path}/completions")
|
||||||
async def completion(request: Request):
|
async def completion(request: Request):
|
||||||
body = await request.body()
|
body = await request.body()
|
||||||
body = body.decode()
|
body_str = body.decode()
|
||||||
data = ast.literal_eval(body)
|
data = ast.literal_eval(body_str)
|
||||||
return litellm_completion(data=data, type="completion", user_model=user_model, user_temperature=user_temperature,
|
return litellm_completion(data=data, type="completion", user_model=user_model, user_temperature=user_temperature,
|
||||||
user_max_tokens=user_max_tokens, user_api_base=user_api_base, user_headers=user_headers,
|
user_max_tokens=user_max_tokens, user_api_base=user_api_base, user_headers=user_headers,
|
||||||
user_debug=user_debug, model_router=model_router, user_request_timeout=user_request_timeout)
|
user_debug=user_debug, model_router=model_router, user_request_timeout=user_request_timeout)
|
||||||
|
@ -519,9 +519,9 @@ async def completion(request: Request):
|
||||||
@router.post("/v1/chat/completions")
|
@router.post("/v1/chat/completions")
|
||||||
@router.post("/chat/completions")
|
@router.post("/chat/completions")
|
||||||
async def chat_completion(request: Request):
|
async def chat_completion(request: Request):
|
||||||
body = body.decode()
|
body = await request.body()
|
||||||
data = ast.literal_eval(body)
|
body_str = body.decode()
|
||||||
print_verbose(f"data passed in: {data}")
|
data = ast.literal_eval(body_str)
|
||||||
return litellm_completion(data, type="chat_completion", user_model=user_model,
|
return litellm_completion(data, type="chat_completion", user_model=user_model,
|
||||||
user_temperature=user_temperature, user_max_tokens=user_max_tokens,
|
user_temperature=user_temperature, user_max_tokens=user_max_tokens,
|
||||||
user_api_base=user_api_base, user_headers=user_headers, user_debug=user_debug, model_router=model_router, user_request_timeout=user_request_timeout)
|
user_api_base=user_api_base, user_headers=user_headers, user_debug=user_debug, model_router=model_router, user_request_timeout=user_request_timeout)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue