litellm/tests
Krish Dholakia a1d9e96b31
LiteLLM Minor Fixes & Improvements (09/25/2024) (#5893)
* fix(langfuse.py): support new langfuse prompt_chat class init params

* fix(langfuse.py): handle new init values on prompt chat + prompt text templates

fixes error caused during langfuse logging

* docs(openai_compatible.md): clarify `openai/` handles correct routing for `/v1/completions` route

Fixes https://github.com/BerriAI/litellm/issues/5876

* fix(utils.py): handle unmapped gemini model optional param translation

Fixes https://github.com/BerriAI/litellm/issues/5888

* fix(o1_transformation.py): fix o-1 validation, to not raise error if temperature=1

Fixes https://github.com/BerriAI/litellm/issues/5884

* fix(prisma_client.py): refresh iam token

Fixes https://github.com/BerriAI/litellm/issues/5896

* fix: pass drop params where required

* fix(utils.py): pass drop_params correctly

* fix(types/vertex_ai.py): fix generation config

* test(test_max_completion_tokens.py): fix test

* fix(vertex_and_google_ai_studio_gemini.py): fix map openai params
2024-09-26 16:41:44 -07:00
..
llm_translation LiteLLM Minor Fixes & Improvements (09/25/2024) (#5893) 2024-09-26 16:41:44 -07:00
load_tests fic otel load test % 2024-09-14 18:04:28 -07:00
otel_tests test(test_otel.py): fix test 2024-09-23 08:10:06 -07:00
pass_through_tests LiteLLM Minor Fixes & Improvements (09/19/2024) (#5793) 2024-09-20 08:19:52 -07:00
proxy_admin_ui_tests [Feat UI sso] store 'provider' in user metadata (#5856) 2024-09-23 17:49:36 -07:00
gettysburg.wav feat(main.py): support openai transcription endpoints 2024-03-08 10:25:19 -08:00
large_text.py fix(router.py): check for context window error when handling 400 status code errors 2024-03-26 08:08:15 -07:00
openai_batch_completions.jsonl feat(router.py): Support Loadbalancing batch azure api endpoints (#5469) 2024-09-02 21:32:55 -07:00
README.MD Update README.MD 2024-03-29 14:56:41 -07:00
test_callbacks_on_proxy.py fix - test num callbacks 2024-05-17 22:06:51 -07:00
test_config.py mark test_team_logging as flaky 2024-09-04 20:29:21 -07:00
test_debug_warning.py fix(utils.py): fix togetherai streaming cost calculation 2024-08-01 15:03:08 -07:00
test_end_users.py test(test_end_users.py): fix test 2024-07-13 21:46:19 -07:00
test_entrypoint.py refactor secret managers 2024-09-03 10:58:02 -07:00
test_fallbacks.py fix(user_api_key_auth.py): ensure user has access to fallback models 2024-06-20 16:02:19 -07:00
test_health.py (test) /health/readiness 2024-01-29 15:27:25 -08:00
test_keys.py mark test_key_info_spend_values_streaming as flaky 2024-08-29 14:39:53 -07:00
test_models.py test - OpenAI client is re-used for Azure, OpenAI 2024-05-10 13:43:19 -07:00
test_openai_batches_endpoint.py test batches endpoint on proxy 2024-07-30 09:46:30 -07:00
test_openai_endpoints.py Litellm fix router testing (#5748) 2024-09-17 18:02:23 -07:00
test_openai_files_endpoints.py test - batches endpoint 2024-07-26 18:09:49 -07:00
test_openai_fine_tuning.py fix cancel ft job route 2024-07-31 16:19:15 -07:00
test_organizations.py test(test_organizations.py): add testing for /organization/new endpoint 2024-03-02 12:13:54 -08:00
test_passthrough_endpoints.py test test_basic_passthrough 2024-08-06 21:17:07 -07:00
test_ratelimit.py test(test_ratelimit.py): fix test to send below rpm 2024-04-30 19:35:21 -07:00
test_spend_logs.py test /spend/report 2024-05-13 15:26:39 -07:00
test_team.py fix(management/utils.py): fix add_member to team when adding user_email 2024-08-10 17:12:09 -07:00
test_team_logging.py mark test_team_logging as flaky 2024-09-04 20:29:21 -07:00
test_users.py refactor(test_users.py): refactor test for user info to use mock endpoints 2024-08-12 18:48:43 -07:00

In total litellm runs 500+ tests Most tests are in /litellm/tests. These are just the tests for the proxy docker image, used for circle ci.