mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix(proxy_server.py): expose flag to disable retries when max parallel request limit is hit
This commit is contained in:
parent
d7b525f391
commit
735fc804ed
2 changed files with 8 additions and 0 deletions
|
@ -2672,6 +2672,13 @@ def giveup(e):
|
|||
and isinstance(e.message, str)
|
||||
and "Max parallel request limit reached" in e.message
|
||||
)
|
||||
|
||||
if (
|
||||
general_settings.get("disable_retry_on_max_parallel_request_limit_error")
|
||||
is True
|
||||
):
|
||||
return True # giveup if queuing max parallel request limits is disabled
|
||||
|
||||
if result:
|
||||
verbose_proxy_logger.info(json.dumps({"event": "giveup", "exception": str(e)}))
|
||||
return result
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue