forked from phoenix/litellm-mirror
* move cooldown logic to it's own helper * add new track deployment metrics folder * increment success, fails for deployment in current minute * fix cooldown logic * fix test_aaarouter_dynamic_cooldown_message_retry_time * fix test_single_deployment_no_cooldowns_test_prod_mock_completion_calls * clean up get from deployment test * fix _async_get_healthy_deployments * add mock InternalServerError * test deployment failing 25% requests * add test_high_traffic_cooldowns_one_bad_deployment * fix vertex load test * add test for rate limit error models in cool down * change default cooldown time * fix cooldown message time * fix cooldown on 429 error * fix doc string for _should_cooldown_deployment * fix sync cooldown logic router |
||
---|---|---|
.. | ||
llm_translation | ||
load_tests | ||
otel_tests | ||
pass_through_tests | ||
proxy_admin_ui_tests | ||
gettysburg.wav | ||
large_text.py | ||
openai_batch_completions.jsonl | ||
README.MD | ||
test_callbacks_on_proxy.py | ||
test_config.py | ||
test_debug_warning.py | ||
test_end_users.py | ||
test_entrypoint.py | ||
test_fallbacks.py | ||
test_health.py | ||
test_keys.py | ||
test_models.py | ||
test_openai_batches_endpoint.py | ||
test_openai_endpoints.py | ||
test_openai_files_endpoints.py | ||
test_openai_fine_tuning.py | ||
test_organizations.py | ||
test_passthrough_endpoints.py | ||
test_ratelimit.py | ||
test_spend_logs.py | ||
test_team.py | ||
test_team_logging.py | ||
test_users.py |
In total litellm runs 500+ tests Most tests are in /litellm/tests. These are just the tests for the proxy docker image, used for circle ci.