mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
fix stream=True
This commit is contained in:
parent
e8b9b4f68b
commit
a5c526a2cd
2 changed files with 11 additions and 1 deletions
|
@ -5,6 +5,9 @@ Handler for transforming responses api requests to litellm.completion requests
|
||||||
from typing import Any, Coroutine, Optional, Union
|
from typing import Any, Coroutine, Optional, Union
|
||||||
|
|
||||||
import litellm
|
import litellm
|
||||||
|
from litellm.responses.litellm_completion_transformation.streaming_iterator import (
|
||||||
|
LiteLLMCompletionStreamingIterator,
|
||||||
|
)
|
||||||
from litellm.responses.litellm_completion_transformation.transformation import (
|
from litellm.responses.litellm_completion_transformation.transformation import (
|
||||||
LiteLLMCompletionResponsesConfig,
|
LiteLLMCompletionResponsesConfig,
|
||||||
)
|
)
|
||||||
|
@ -26,6 +29,7 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
responses_api_request: ResponsesAPIOptionalRequestParams,
|
responses_api_request: ResponsesAPIOptionalRequestParams,
|
||||||
custom_llm_provider: Optional[str] = None,
|
custom_llm_provider: Optional[str] = None,
|
||||||
_is_async: bool = False,
|
_is_async: bool = False,
|
||||||
|
stream: Optional[bool] = None,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
) -> Union[
|
) -> Union[
|
||||||
ResponsesAPIResponse,
|
ResponsesAPIResponse,
|
||||||
|
@ -40,6 +44,8 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
input=input,
|
input=input,
|
||||||
responses_api_request=responses_api_request,
|
responses_api_request=responses_api_request,
|
||||||
custom_llm_provider=custom_llm_provider,
|
custom_llm_provider=custom_llm_provider,
|
||||||
|
stream=stream,
|
||||||
|
**kwargs,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -88,4 +94,7 @@ class LiteLLMCompletionTransformationHandler:
|
||||||
|
|
||||||
return responses_api_response
|
return responses_api_response
|
||||||
|
|
||||||
raise ValueError("litellm_completion_response is not a ModelResponse")
|
elif isinstance(litellm_completion_response, litellm.CustomStreamWrapper):
|
||||||
|
return LiteLLMCompletionStreamingIterator(
|
||||||
|
litellm_custom_stream_wrapper=litellm_completion_response,
|
||||||
|
)
|
||||||
|
|
|
@ -204,6 +204,7 @@ def responses(
|
||||||
responses_api_request=ResponsesAPIOptionalRequestParams(),
|
responses_api_request=ResponsesAPIOptionalRequestParams(),
|
||||||
custom_llm_provider=custom_llm_provider,
|
custom_llm_provider=custom_llm_provider,
|
||||||
_is_async=_is_async,
|
_is_async=_is_async,
|
||||||
|
stream=stream,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue