mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
fix(utils.py): fix print statement
This commit is contained in:
parent
45eb4a5fcc
commit
7b641491a2
1 changed files with 1 additions and 3 deletions
|
@ -8752,9 +8752,7 @@ class CustomStreamWrapper:
|
|||
or self.custom_llm_provider in litellm.openai_compatible_endpoints
|
||||
):
|
||||
async for chunk in self.completion_stream:
|
||||
print_verbose(
|
||||
f"value of async chunk: {chunk.parts}; len(chunk.parts): {len(chunk.parts)}"
|
||||
)
|
||||
print_verbose(f"value of async chunk: {chunk}")
|
||||
if chunk == "None" or chunk is None:
|
||||
raise Exception
|
||||
elif self.custom_llm_provider == "gemini" and len(chunk.parts) == 0:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue