Merge pull request #2203 from BerriAI/litellm_streaming_caching_fix

fix(utils.py): support returning caching streaming response for function calling streaming calls
This commit is contained in:
Krish Dholakia 2024-02-26 19:58:00 -08:00 committed by GitHub
commit 95b5b7f1fc
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
5 changed files with 168 additions and 64 deletions

View file

@ -3677,6 +3677,7 @@ def stream_chunk_builder(
response["usage"]["total_tokens"] = (
response["usage"]["prompt_tokens"] + response["usage"]["completion_tokens"]
)
return convert_to_model_response_object(
response_object=response,
model_response_object=model_response,