forked from phoenix/litellm-mirror
(fix) proxy raise exceptions from litellm.completion()
This commit is contained in:
parent
e698f8218f
commit
a18fe5a1fd
1 changed files with 13 additions and 10 deletions
|
@ -455,16 +455,19 @@ def litellm_completion(*args, **kwargs):
|
||||||
kwargs["api_base"] = user_api_base
|
kwargs["api_base"] = user_api_base
|
||||||
## ROUTE TO CORRECT ENDPOINT ##
|
## ROUTE TO CORRECT ENDPOINT ##
|
||||||
router_model_names = [m["model_name"] for m in llm_model_list] if llm_model_list is not None else []
|
router_model_names = [m["model_name"] for m in llm_model_list] if llm_model_list is not None else []
|
||||||
if llm_router is not None and kwargs["model"] in router_model_names: # model in router model list
|
try:
|
||||||
if call_type == "chat_completion":
|
if llm_router is not None and kwargs["model"] in router_model_names: # model in router model list
|
||||||
response = llm_router.completion(*args, **kwargs)
|
if call_type == "chat_completion":
|
||||||
elif call_type == "text_completion":
|
response = llm_router.completion(*args, **kwargs)
|
||||||
response = llm_router.text_completion(*args, **kwargs)
|
elif call_type == "text_completion":
|
||||||
else:
|
response = llm_router.text_completion(*args, **kwargs)
|
||||||
if call_type == "chat_completion":
|
else:
|
||||||
response = litellm.completion(*args, **kwargs)
|
if call_type == "chat_completion":
|
||||||
elif call_type == "text_completion":
|
response = litellm.completion(*args, **kwargs)
|
||||||
response = litellm.text_completion(*args, **kwargs)
|
elif call_type == "text_completion":
|
||||||
|
response = litellm.text_completion(*args, **kwargs)
|
||||||
|
except Exception as e:
|
||||||
|
raise e
|
||||||
if 'stream' in kwargs and kwargs['stream'] == True: # use generate_responses to stream responses
|
if 'stream' in kwargs and kwargs['stream'] == True: # use generate_responses to stream responses
|
||||||
return StreamingResponse(data_generator(response), media_type='text/event-stream')
|
return StreamingResponse(data_generator(response), media_type='text/event-stream')
|
||||||
return response
|
return response
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue