diff --git a/litellm/proxy/_types.py b/litellm/proxy/_types.py index 25aa942e5..d3f1bc844 100644 --- a/litellm/proxy/_types.py +++ b/litellm/proxy/_types.py @@ -1144,6 +1144,14 @@ class ConfigGeneralSettings(LiteLLMBase): global_max_parallel_requests: Optional[int] = Field( None, description="global max parallel requests to allow for a proxy instance." ) + max_request_size_mb: Optional[int] = Field( + None, + description="max request size in MB, if a request is larger than this size it will be rejected", + ) + max_response_size_mb: Optional[int] = Field( + None, + description="max response size in MB, if a response is larger than this size it will be rejected", + ) infer_model_from_keys: Optional[bool] = Field( None, description="for `/models` endpoint, infers available model based on environment keys (e.g. OPENAI_API_KEY)", diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index 46f67b610..33f021210 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -9083,6 +9083,8 @@ async def get_config_list( allowed_args = { "max_parallel_requests": {"type": "Integer"}, "global_max_parallel_requests": {"type": "Integer"}, + "max_request_size_mb": {"type": "Integer"}, + "max_response_size_mb": {"type": "Integer"}, } return_val = []