diff --git a/litellm/tests/test_mock_request.py b/litellm/tests/test_mock_request.py index 1997d454d..82eb30926 100644 --- a/litellm/tests/test_mock_request.py +++ b/litellm/tests/test_mock_request.py @@ -27,7 +27,6 @@ def test_streaming_mock_request(): response = litellm.mock_completion(model=model, messages=messages, stream=True) complete_response = "" for chunk in response: - print(f"chunk: {chunk}") complete_response += chunk["choices"][0]["delta"]["content"] if complete_response == "": raise Exception("Empty response received") diff --git a/litellm/utils.py b/litellm/utils.py index 2a1cbe0c2..13411c13c 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -2922,7 +2922,6 @@ class CustomStreamWrapper: raise ValueError(f"Unable to parse response. Original response: {chunk}") def handle_replicate_chunk(self, chunk): - print(f"chunk: {chunk}") try: text = "" is_finished = False @@ -2941,7 +2940,6 @@ class CustomStreamWrapper: def handle_openai_text_completion_chunk(self, chunk): try: - print(f"chunk: {chunk}") return chunk["choices"][0]["text"] except: raise ValueError(f"Unable to parse response. Original response: {chunk}") diff --git a/pyproject.toml b/pyproject.toml index db07be1ed..bff28dcce 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.poetry] name = "litellm" -version = "0.1.784" +version = "0.1.785" description = "Library to easily interface with LLM API providers" authors = ["BerriAI"] license = "MIT License"