forked from phoenix/litellm-mirror
LiteLLM Minor Fixes and Improvements (11/09/2024) (#5634)
* fix(caching.py): set ttl for async_increment cache fixes issue where ttl for redis client was not being set on increment_cache Fixes https://github.com/BerriAI/litellm/issues/5609 * fix(caching.py): fix increment cache w/ ttl for sync increment cache on redis Fixes https://github.com/BerriAI/litellm/issues/5609 * fix(router.py): support adding retry policy + allowed fails policy via config.yaml * fix(router.py): don't cooldown single deployments No point, as there's no other deployment to loadbalance with. * fix(user_api_key_auth.py): support setting allowed email domains on jwt tokens Closes https://github.com/BerriAI/litellm/issues/5605 * docs(token_auth.md): add user upsert + allowed email domain to jwt auth docs * fix(litellm_pre_call_utils.py): fix dynamic key logging when team id is set Fixes issue where key logging would not be set if team metadata was not none * fix(secret_managers/main.py): load environment variables correctly Fixes issue where os.environ/ was not being loaded correctly * test(test_router.py): fix test * feat(spend_tracking_utils.py): support logging additional usage params - e.g. prompt caching values for deepseek * test: fix tests * test: fix test * test: fix test * test: fix test * test: fix test
This commit is contained in:
parent
70100d716b
commit
98c34a7e27
25 changed files with 745 additions and 114 deletions
|
@ -1609,15 +1609,24 @@ class Logging:
|
|||
"""
|
||||
from litellm.types.router import RouterErrors
|
||||
|
||||
litellm_params: dict = self.model_call_details.get("litellm_params") or {}
|
||||
metadata = litellm_params.get("metadata") or {}
|
||||
|
||||
## BASE CASE ## check if rate limit error for model group size 1
|
||||
is_base_case = False
|
||||
if metadata.get("model_group_size") is not None:
|
||||
model_group_size = metadata.get("model_group_size")
|
||||
if isinstance(model_group_size, int) and model_group_size == 1:
|
||||
is_base_case = True
|
||||
## check if special error ##
|
||||
if RouterErrors.no_deployments_available.value not in str(exception):
|
||||
if (
|
||||
RouterErrors.no_deployments_available.value not in str(exception)
|
||||
and is_base_case is False
|
||||
):
|
||||
return
|
||||
|
||||
## get original model group ##
|
||||
|
||||
litellm_params: dict = self.model_call_details.get("litellm_params") or {}
|
||||
metadata = litellm_params.get("metadata") or {}
|
||||
|
||||
model_group = metadata.get("model_group") or None
|
||||
for callback in litellm._async_failure_callback:
|
||||
if isinstance(callback, CustomLogger): # custom logger class
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue