litellm-mirror/litellm/router_utils
Krish Dholakia 513b1904ab
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 14s
Add attempted-retries and timeout values to response headers + more testing (#7926)
* feat(router.py): add retry headers to response

makes it easy to add testing to ensure model-specific retries are respected

* fix(add_retry_headers.py): clarify attempted retries vs. max retries

* test(test_fallbacks.py): add test for checking if max retries set for model is respected

* test(test_fallbacks.py): assert values for attempted retries and max retries are as expected

* fix(utils.py): return timeout in litellm proxy response headers

* test(test_fallbacks.py): add test to assert model specific timeout used on timeout error

* test: add bad model with timeout to proxy

* fix: fix linting error

* fix(router.py): fix get model list from model alias

* test: loosen test restriction - account for other events on proxy
2025-01-22 22:19:44 -08:00
..
pre_call_checks Litellm dev readd prompt caching (#7299) 2024-12-18 15:13:49 -08:00
router_callbacks (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
add_retry_headers.py Add attempted-retries and timeout values to response headers + more testing (#7926) 2025-01-22 22:19:44 -08:00
batch_utils.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
client_initalization_utils.py (Feat) - LiteLLM Use UsernamePasswordCredential for Azure OpenAI (#7496) 2025-01-01 14:11:27 -08:00
cooldown_cache.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
cooldown_callbacks.py Litellm dev readd prompt caching (#7299) 2024-12-18 15:13:49 -08:00
cooldown_handlers.py Improve Proxy Resiliency: Cooldown single-deployment model groups if 100% calls failed in high traffic (#7823) 2025-01-17 20:17:02 -08:00
fallback_event_handlers.py Controll fallback prompts client-side (#7334) 2024-12-20 19:09:53 -08:00
get_retry_from_policy.py Litellm dev 12 06 2024 (#7067) 2024-12-06 22:44:18 -08:00
handle_error.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
pattern_match_deployments.py Litellm dev 12 28 2024 p2 (#7458) 2024-12-28 19:38:06 -08:00
prompt_caching_cache.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
response_headers.py LiteLLM Minor Fixes & Improvements (11/26/2024) (#6913) 2024-11-28 00:01:38 +05:30