mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
bump: version 1.7.13 → 1.7.14
This commit is contained in:
parent
451851e6a4
commit
6c98715b94
3 changed files with 4 additions and 2 deletions
|
@ -2212,6 +2212,7 @@ def stream_chunk_builder(chunks: list, messages: Optional[list]=None):
|
|||
try:
|
||||
response["usage"]["prompt_tokens"] = token_counter(model=model, messages=messages)
|
||||
except: # don't allow this failing to block a complete streaming response from being returned
|
||||
print_verbose(f"token_counter failed, assuming prompt tokens is 0")
|
||||
response["usage"]["prompt_tokens"] = 0
|
||||
response["usage"]["completion_tokens"] = token_counter(model=model, text=completion_output)
|
||||
response["usage"]["total_tokens"] = response["usage"]["prompt_tokens"] + response["usage"]["completion_tokens"]
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue