mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
fix(utils.py): fix cost calculation to handle tool input
This commit is contained in:
parent
9494c2cd9e
commit
1b35736797
3 changed files with 77 additions and 6 deletions
|
@ -1858,8 +1858,6 @@ def completion_cost(
|
|||
- If an error occurs during execution, the function returns 0.0 without blocking the user's execution path.
|
||||
"""
|
||||
try:
|
||||
if messages != []:
|
||||
prompt = " ".join([message["content"] for message in messages])
|
||||
# Handle Inputs to completion_cost
|
||||
prompt_tokens = 0
|
||||
completion_tokens = 0
|
||||
|
@ -1869,7 +1867,10 @@ def completion_cost(
|
|||
completion_tokens = completion_response['usage']['completion_tokens']
|
||||
model = model or completion_response['model'] # check if user passed an override for model, if it's none check completion_response['model']
|
||||
else:
|
||||
prompt_tokens = token_counter(model=model, text=prompt)
|
||||
if len(messages) > 0:
|
||||
prompt_tokens = token_counter(model=model, messages=messages)
|
||||
elif len(prompt) > 0:
|
||||
prompt_tokens = token_counter(model=model, text=prompt)
|
||||
completion_tokens = token_counter(model=model, text=completion)
|
||||
|
||||
# Calculate cost based on prompt_tokens, completion_tokens
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue