mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
fix handler
This commit is contained in:
parent
e91bdff9f4
commit
e29219e55c
2 changed files with 16 additions and 0 deletions
|
@ -5,3 +5,9 @@ model_list:
|
||||||
- model_name: anthropic/*
|
- model_name: anthropic/*
|
||||||
litellm_params:
|
litellm_params:
|
||||||
model: anthropic/*
|
model: anthropic/*
|
||||||
|
- model_name: gemini/*
|
||||||
|
litellm_params:
|
||||||
|
model: gemini/*
|
||||||
|
litellm_settings:
|
||||||
|
drop_params: true
|
||||||
|
|
||||||
|
|
|
@ -52,6 +52,8 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
if _is_async:
|
if _is_async:
|
||||||
return self.async_response_api_handler(
|
return self.async_response_api_handler(
|
||||||
litellm_completion_request=litellm_completion_request,
|
litellm_completion_request=litellm_completion_request,
|
||||||
|
request_input=input,
|
||||||
|
responses_api_request=responses_api_request,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -66,6 +68,8 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
responses_api_response: ResponsesAPIResponse = (
|
responses_api_response: ResponsesAPIResponse = (
|
||||||
LiteLLMCompletionResponsesConfig.transform_chat_completion_response_to_responses_api_response(
|
LiteLLMCompletionResponsesConfig.transform_chat_completion_response_to_responses_api_response(
|
||||||
chat_completion_response=litellm_completion_response,
|
chat_completion_response=litellm_completion_response,
|
||||||
|
request_input=input,
|
||||||
|
responses_api_request=responses_api_request,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -76,6 +80,8 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
async def async_response_api_handler(
|
async def async_response_api_handler(
|
||||||
self,
|
self,
|
||||||
litellm_completion_request: dict,
|
litellm_completion_request: dict,
|
||||||
|
request_input: Union[str, ResponseInputParam],
|
||||||
|
responses_api_request: ResponsesAPIOptionalRequestParams,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
) -> Union[ResponsesAPIResponse, BaseResponsesAPIStreamingIterator]:
|
) -> Union[ResponsesAPIResponse, BaseResponsesAPIStreamingIterator]:
|
||||||
litellm_completion_response: Union[
|
litellm_completion_response: Union[
|
||||||
|
@ -89,6 +95,8 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
responses_api_response: ResponsesAPIResponse = (
|
responses_api_response: ResponsesAPIResponse = (
|
||||||
LiteLLMCompletionResponsesConfig.transform_chat_completion_response_to_responses_api_response(
|
LiteLLMCompletionResponsesConfig.transform_chat_completion_response_to_responses_api_response(
|
||||||
chat_completion_response=litellm_completion_response,
|
chat_completion_response=litellm_completion_response,
|
||||||
|
request_input=request_input,
|
||||||
|
responses_api_request=responses_api_request,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -97,4 +105,6 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
elif isinstance(litellm_completion_response, litellm.CustomStreamWrapper):
|
elif isinstance(litellm_completion_response, litellm.CustomStreamWrapper):
|
||||||
return LiteLLMCompletionStreamingIterator(
|
return LiteLLMCompletionStreamingIterator(
|
||||||
litellm_custom_stream_wrapper=litellm_completion_response,
|
litellm_custom_stream_wrapper=litellm_completion_response,
|
||||||
|
request_input=request_input,
|
||||||
|
responses_api_request=responses_api_request,
|
||||||
)
|
)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue