mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
fix(utils.py): fix print statement
This commit is contained in:
parent
45eb4a5fcc
commit
7b641491a2
1 changed files with 1 additions and 3 deletions
|
@ -8752,9 +8752,7 @@ class CustomStreamWrapper:
|
||||||
or self.custom_llm_provider in litellm.openai_compatible_endpoints
|
or self.custom_llm_provider in litellm.openai_compatible_endpoints
|
||||||
):
|
):
|
||||||
async for chunk in self.completion_stream:
|
async for chunk in self.completion_stream:
|
||||||
print_verbose(
|
print_verbose(f"value of async chunk: {chunk}")
|
||||||
f"value of async chunk: {chunk.parts}; len(chunk.parts): {len(chunk.parts)}"
|
|
||||||
)
|
|
||||||
if chunk == "None" or chunk is None:
|
if chunk == "None" or chunk is None:
|
||||||
raise Exception
|
raise Exception
|
||||||
elif self.custom_llm_provider == "gemini" and len(chunk.parts) == 0:
|
elif self.custom_llm_provider == "gemini" and len(chunk.parts) == 0:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue