mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
fix - undo local dev changes
This commit is contained in:
parent
6b59aeb603
commit
dfc22194b2
1 changed files with 3 additions and 3 deletions
|
@ -1707,7 +1707,7 @@ class Router:
|
|||
response = original_function(*args, **kwargs)
|
||||
return response
|
||||
|
||||
except Exception as most_recent_exception:
|
||||
except Exception as e:
|
||||
## LOGGING
|
||||
kwargs = self.log_retry(kwargs=kwargs, e=e)
|
||||
remaining_retries = num_retries - current_attempt
|
||||
|
@ -1717,7 +1717,7 @@ class Router:
|
|||
num_retries=num_retries,
|
||||
)
|
||||
time.sleep(_timeout)
|
||||
raise most_recent_exception
|
||||
raise original_exception
|
||||
|
||||
### HELPER FUNCTIONS
|
||||
|
||||
|
@ -1852,7 +1852,7 @@ class Router:
|
|||
exception_status = 500
|
||||
_should_retry = litellm._should_retry(status_code=exception_status)
|
||||
|
||||
if updated_fails > self.allowed_fails:
|
||||
if updated_fails > self.allowed_fails or _should_retry == False:
|
||||
# get the current cooldown list for that minute
|
||||
cooldown_key = f"{current_minute}:cooldown_models" # group cooldown models by minute to reduce number of redis calls
|
||||
cached_value = self.cache.get_cache(key=cooldown_key)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue