From 1f51159ed20278f23a3d7bfe2d67fd3a8aed5e8a Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Fri, 27 Sep 2024 22:58:48 -0700 Subject: [PATCH] =?UTF-8?q?bump:=20version=201.48.4=20=E2=86=92=201.48.5?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- litellm/__init__.py | 2 +- pyproject.toml | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/litellm/__init__.py b/litellm/__init__.py index d76dd37bc..c7f290754 100644 --- a/litellm/__init__.py +++ b/litellm/__init__.py @@ -269,7 +269,7 @@ max_end_user_budget: Optional[float] = None priority_reservation: Optional[Dict[str, float]] = None #### RELIABILITY #### REPEATED_STREAMING_CHUNK_LIMIT = 100 # catch if model starts looping the same chunk while streaming. Uses high default to prevent false positives. -request_timeout: float = 6000 +request_timeout: float = 6000 # time in seconds module_level_aclient = AsyncHTTPHandler(timeout=request_timeout) module_level_client = HTTPHandler(timeout=request_timeout) num_retries: Optional[int] = None # per model endpoint diff --git a/pyproject.toml b/pyproject.toml index fc1a3438a..08556eff6 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.poetry] name = "litellm" -version = "1.48.4" +version = "1.48.5" description = "Library to easily interface with LLM API providers" authors = ["BerriAI"] license = "MIT" @@ -91,7 +91,7 @@ requires = ["poetry-core", "wheel"] build-backend = "poetry.core.masonry.api" [tool.commitizen] -version = "1.48.4" +version = "1.48.5" version_files = [ "pyproject.toml:^version" ]