diff --git a/litellm/__init__.py b/litellm/__init__.py index d76dd37bc..c7f290754 100644 --- a/litellm/__init__.py +++ b/litellm/__init__.py @@ -269,7 +269,7 @@ max_end_user_budget: Optional[float] = None priority_reservation: Optional[Dict[str, float]] = None #### RELIABILITY #### REPEATED_STREAMING_CHUNK_LIMIT = 100 # catch if model starts looping the same chunk while streaming. Uses high default to prevent false positives. -request_timeout: float = 6000 +request_timeout: float = 6000 # time in seconds module_level_aclient = AsyncHTTPHandler(timeout=request_timeout) module_level_client = HTTPHandler(timeout=request_timeout) num_retries: Optional[int] = None # per model endpoint diff --git a/pyproject.toml b/pyproject.toml index fc1a3438a..08556eff6 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.poetry] name = "litellm" -version = "1.48.4" +version = "1.48.5" description = "Library to easily interface with LLM API providers" authors = ["BerriAI"] license = "MIT" @@ -91,7 +91,7 @@ requires = ["poetry-core", "wheel"] build-backend = "poetry.core.masonry.api" [tool.commitizen] -version = "1.48.4" +version = "1.48.5" version_files = [ "pyproject.toml:^version" ]