mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
Fix completion_tokens on Vertex AI Gemini thinking models
This commit is contained in:
parent
62ed5b2f78
commit
f87f500a49
1 changed files with 9 additions and 3 deletions
|
@ -766,14 +766,20 @@ class VertexGeminiConfig(VertexAIBaseConfig, BaseConfig):
|
|||
audio_tokens=audio_tokens,
|
||||
text_tokens=text_tokens,
|
||||
)
|
||||
completion_tokens = completion_response["usageMetadata"].get(
|
||||
"candidatesTokenCount", 0
|
||||
)
|
||||
if reasoning_tokens:
|
||||
# Usage(...) constructor expects that completion_tokens includes the reasoning_tokens.
|
||||
# However the Vertex AI usage metadata does not include reasoning tokens in candidatesTokenCount.
|
||||
# Reportedly, this is different from the Gemini API.
|
||||
completion_tokens += reasoning_tokens
|
||||
## GET USAGE ##
|
||||
usage = Usage(
|
||||
prompt_tokens=completion_response["usageMetadata"].get(
|
||||
"promptTokenCount", 0
|
||||
),
|
||||
completion_tokens=completion_response["usageMetadata"].get(
|
||||
"candidatesTokenCount", 0
|
||||
),
|
||||
completion_tokens=completion_tokens,
|
||||
total_tokens=completion_response["usageMetadata"].get("totalTokenCount", 0),
|
||||
prompt_tokens_details=prompt_tokens_details,
|
||||
reasoning_tokens=reasoning_tokens,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue