From 7b641491a2fb1a20912f002bf67f490b6aac9d1a Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Mon, 19 Feb 2024 23:00:41 -0800 Subject: [PATCH] fix(utils.py): fix print statement --- litellm/utils.py | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/litellm/utils.py b/litellm/utils.py index c299a440d..982462e3f 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -8752,9 +8752,7 @@ class CustomStreamWrapper: or self.custom_llm_provider in litellm.openai_compatible_endpoints ): async for chunk in self.completion_stream: - print_verbose( - f"value of async chunk: {chunk.parts}; len(chunk.parts): {len(chunk.parts)}" - ) + print_verbose(f"value of async chunk: {chunk}") if chunk == "None" or chunk is None: raise Exception elif self.custom_llm_provider == "gemini" and len(chunk.parts) == 0: