diff --git a/dist/litellm-1.12.6.dev2-py3-none-any.whl b/dist/litellm-1.12.6.dev2-py3-none-any.whl new file mode 100644 index 000000000..bf6d294b9 Binary files /dev/null and b/dist/litellm-1.12.6.dev2-py3-none-any.whl differ diff --git a/dist/litellm-1.12.6.dev2.tar.gz b/dist/litellm-1.12.6.dev2.tar.gz new file mode 100644 index 000000000..6f032ba7b Binary files /dev/null and b/dist/litellm-1.12.6.dev2.tar.gz differ diff --git a/litellm/main.py b/litellm/main.py index e12b64187..d90e1c3f2 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -202,10 +202,12 @@ async def acompletion(*args, **kwargs): async def _async_streaming(response, model, custom_llm_provider, args): try: + print_verbose(f"received response in _async_streaming: {response}") async for line in response: print_verbose(f"line in async streaming: {line}") yield line except Exception as e: + print_verbose(f"error raised _async_streaming: {str(e)}") raise exception_type( model=model, custom_llm_provider=custom_llm_provider, original_exception=e, completion_kwargs=args, ) diff --git a/litellm/utils.py b/litellm/utils.py index 55c7d76c3..8f4f76db8 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -5688,6 +5688,7 @@ class CustomStreamWrapper: chunk = self.completion_stream else: chunk = next(self.completion_stream) + print_verbose(f"value of chunk: {chunk} ") if chunk is not None and chunk != b'': print_verbose(f"PROCESSED CHUNK PRE CHUNK CREATOR: {chunk}") response = self.chunk_creator(chunk=chunk) @@ -5728,6 +5729,7 @@ class CustomStreamWrapper: raise StopAsyncIteration else: # temporary patch for non-aiohttp async calls # example - boto3 bedrock llms + print_verbose(f"ENTERS __NEXT__ LOOP") processed_chunk = next(self) asyncio.create_task(self.logging_obj.async_success_handler(processed_chunk,)) print_verbose(f"PROCESSED CHUNK IN __ANEXT__: {processed_chunk}")