From dacadbf624d83d0e248ddf7a404c136c0dd7b044 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Wed, 24 Apr 2024 20:56:10 -0700 Subject: [PATCH] fix(utils.py): fix anthropic streaming return usage tokens --- litellm/utils.py | 1 + 1 file changed, 1 insertion(+) diff --git a/litellm/utils.py b/litellm/utils.py index 9b87331575..0c0b6b1777 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -9758,6 +9758,7 @@ class CustomStreamWrapper: if response_obj is None: return completion_obj["content"] = response_obj["text"] + setattr(model_response, "usage", Usage()) if response_obj.get("prompt_tokens", None) is not None: model_response.usage.prompt_tokens = response_obj[ "prompt_tokens"