mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
allow setting max request / response size on admin UI
This commit is contained in:
parent
f633f7d92d
commit
3511aadf99
2 changed files with 10 additions and 0 deletions
|
@ -1144,6 +1144,14 @@ class ConfigGeneralSettings(LiteLLMBase):
|
||||||
global_max_parallel_requests: Optional[int] = Field(
|
global_max_parallel_requests: Optional[int] = Field(
|
||||||
None, description="global max parallel requests to allow for a proxy instance."
|
None, description="global max parallel requests to allow for a proxy instance."
|
||||||
)
|
)
|
||||||
|
max_request_size_mb: Optional[int] = Field(
|
||||||
|
None,
|
||||||
|
description="max request size in MB, if a request is larger than this size it will be rejected",
|
||||||
|
)
|
||||||
|
max_response_size_mb: Optional[int] = Field(
|
||||||
|
None,
|
||||||
|
description="max response size in MB, if a response is larger than this size it will be rejected",
|
||||||
|
)
|
||||||
infer_model_from_keys: Optional[bool] = Field(
|
infer_model_from_keys: Optional[bool] = Field(
|
||||||
None,
|
None,
|
||||||
description="for `/models` endpoint, infers available model based on environment keys (e.g. OPENAI_API_KEY)",
|
description="for `/models` endpoint, infers available model based on environment keys (e.g. OPENAI_API_KEY)",
|
||||||
|
|
|
@ -9083,6 +9083,8 @@ async def get_config_list(
|
||||||
allowed_args = {
|
allowed_args = {
|
||||||
"max_parallel_requests": {"type": "Integer"},
|
"max_parallel_requests": {"type": "Integer"},
|
||||||
"global_max_parallel_requests": {"type": "Integer"},
|
"global_max_parallel_requests": {"type": "Integer"},
|
||||||
|
"max_request_size_mb": {"type": "Integer"},
|
||||||
|
"max_response_size_mb": {"type": "Integer"},
|
||||||
}
|
}
|
||||||
|
|
||||||
return_val = []
|
return_val = []
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue