forked from phoenix/litellm-mirror
* fix(model_prices_and_context_window.json): add cost tracking for more vertex llama3.1 model 8b and 70b models * fix(proxy/utils.py): handle data being none on pre-call hooks * fix(proxy/): create views on initial proxy startup fixes base case, where user starts proxy for first time Fixes https://github.com/BerriAI/litellm/issues/5756 * build(config.yml): fix vertex version for test * feat(ui/): support enabling/disabling slack alerting Allows admin to turn on/off slack alerting through ui * feat(rerank/main.py): support langfuse logging * fix(proxy/utils.py): fix linting errors * fix(langfuse.py): log clean metadata * test(tests): replace deprecated openai model |
||
---|---|---|
.. | ||
llm_translation | ||
load_tests | ||
otel_tests | ||
pass_through_tests | ||
proxy_admin_ui_tests | ||
gettysburg.wav | ||
large_text.py | ||
openai_batch_completions.jsonl | ||
README.MD | ||
test_callbacks_on_proxy.py | ||
test_config.py | ||
test_debug_warning.py | ||
test_end_users.py | ||
test_entrypoint.py | ||
test_fallbacks.py | ||
test_health.py | ||
test_keys.py | ||
test_models.py | ||
test_openai_batches_endpoint.py | ||
test_openai_endpoints.py | ||
test_openai_files_endpoints.py | ||
test_openai_fine_tuning.py | ||
test_organizations.py | ||
test_passthrough_endpoints.py | ||
test_ratelimit.py | ||
test_spend_logs.py | ||
test_team.py | ||
test_team_logging.py | ||
test_users.py |
In total litellm runs 500+ tests Most tests are in /litellm/tests. These are just the tests for the proxy docker image, used for circle ci.