mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
bump: version 1.7.13 → 1.7.14
This commit is contained in:
parent
451851e6a4
commit
6c98715b94
3 changed files with 4 additions and 2 deletions
|
@ -2212,6 +2212,7 @@ def stream_chunk_builder(chunks: list, messages: Optional[list]=None):
|
||||||
try:
|
try:
|
||||||
response["usage"]["prompt_tokens"] = token_counter(model=model, messages=messages)
|
response["usage"]["prompt_tokens"] = token_counter(model=model, messages=messages)
|
||||||
except: # don't allow this failing to block a complete streaming response from being returned
|
except: # don't allow this failing to block a complete streaming response from being returned
|
||||||
|
print_verbose(f"token_counter failed, assuming prompt tokens is 0")
|
||||||
response["usage"]["prompt_tokens"] = 0
|
response["usage"]["prompt_tokens"] = 0
|
||||||
response["usage"]["completion_tokens"] = token_counter(model=model, text=completion_output)
|
response["usage"]["completion_tokens"] = token_counter(model=model, text=completion_output)
|
||||||
response["usage"]["total_tokens"] = response["usage"]["prompt_tokens"] + response["usage"]["completion_tokens"]
|
response["usage"]["total_tokens"] = response["usage"]["prompt_tokens"] + response["usage"]["completion_tokens"]
|
||||||
|
|
|
@ -1613,6 +1613,7 @@ def token_counter(model="", text=None, messages: Optional[List] = None):
|
||||||
# use tiktoken, anthropic, cohere or llama2's tokenizer depending on the model
|
# use tiktoken, anthropic, cohere or llama2's tokenizer depending on the model
|
||||||
if text == None:
|
if text == None:
|
||||||
if messages is not None:
|
if messages is not None:
|
||||||
|
print_verbose(f"token_counter messages received: {messages}")
|
||||||
text = "".join([message["content"] for message in messages])
|
text = "".join([message["content"] for message in messages])
|
||||||
else:
|
else:
|
||||||
raise ValueError("text and messages cannot both be None")
|
raise ValueError("text and messages cannot both be None")
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
[tool.poetry]
|
[tool.poetry]
|
||||||
name = "litellm"
|
name = "litellm"
|
||||||
version = "1.7.13"
|
version = "1.7.14"
|
||||||
description = "Library to easily interface with LLM API providers"
|
description = "Library to easily interface with LLM API providers"
|
||||||
authors = ["BerriAI"]
|
authors = ["BerriAI"]
|
||||||
license = "MIT License"
|
license = "MIT License"
|
||||||
|
@ -27,7 +27,7 @@ requires = ["poetry-core"]
|
||||||
build-backend = "poetry.core.masonry.api"
|
build-backend = "poetry.core.masonry.api"
|
||||||
|
|
||||||
[tool.commitizen]
|
[tool.commitizen]
|
||||||
version = "1.7.13"
|
version = "1.7.14"
|
||||||
version_files = [
|
version_files = [
|
||||||
"pyproject.toml:^version"
|
"pyproject.toml:^version"
|
||||||
]
|
]
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue