Commit graph

13391 commits

Author SHA1 Message Date
Krrish Dholakia
db92956ae3 fix(redis_cache.py): add 5s default timeout 2025-03-17 14:27:36 -07:00
Krish Dholakia
d0d8ec2c40
Merge branch 'main' into litellm_dev_03_16_2025_p1 2025-03-17 10:02:53 -07:00
Krrish Dholakia
a5b497667c fix(logging_utils.py): revert change 2025-03-16 21:04:41 -07:00
Krrish Dholakia
a99251a4ab fix(streaming_handler.py): raise stop iteration post-finish reason 2025-03-16 20:40:41 -07:00
Krrish Dholakia
bde9ae8a95 fix(litellm_logging.py): remove unused import 2025-03-16 20:24:27 -07:00
Krrish Dholakia
c0a76427d2 fix(streaming_handler.py): pass complete streaming response on completion 2025-03-16 20:22:12 -07:00
Krrish Dholakia
08b297230e fix(streaming_handler.py): return model response on finished chunk 2025-03-16 13:05:46 -07:00
Krrish Dholakia
b093157369 fix(converse_transformation.py): fix linting error 2025-03-15 19:33:17 -07:00
Krrish Dholakia
5dc46f0cf7 fix(converse_transformation.py): fix encoding model 2025-03-15 14:03:37 -07:00
Krrish Dholakia
612d5a284d refactor(litellm_logging.py): delegate returning a complete response to the streaming_handler
Removes incorrect logic for calculating complete streaming response from litellm logging
2025-03-15 09:55:33 -07:00
Krrish Dholakia
dd2c980d5b fix(utils.py): Prevents final chunk w/ usage from being ignored
Fixes https://github.com/BerriAI/litellm/issues/7112
2025-03-15 09:12:14 -07:00
Krish Dholakia
d4caaae1be
Merge pull request #9274 from BerriAI/litellm_contributor_rebase_branch
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 43s
Helm unit test / unit-test (push) Successful in 50s
Litellm contributor rebase branch
2025-03-14 21:57:49 -07:00
Ishaan Jaff
a256108f3f ui new build 2025-03-14 21:46:40 -07:00
Ishaan Jaff
9dc962511f ui new build 2025-03-14 21:29:42 -07:00
Ishaan Jaff
5290d141b7 ui new build 2025-03-14 21:23:13 -07:00
Ishaan Jaff
12188f286c
Merge pull request #9258 from BerriAI/litellm_fix_models_view_edit
(UI) Fix model edit + delete - instantly show edit + deletes to models
2025-03-14 21:21:46 -07:00
Ishaan Jaff
bda5fe0fcf
Merge pull request #9272 from BerriAI/litellm_add_test_connection_button
[Feat] UI - Add Test Connection
2025-03-14 21:16:44 -07:00
Krrish Dholakia
a9dceacc1b fix(factory.py): reduce ollama pt LOC < 50 2025-03-14 21:10:05 -07:00
Ishaan Jaff
d7e10fee79 fix code quality 2025-03-14 21:06:28 -07:00
Ishaan Jaff
cbf0fa44b4 undo changes to route llm request 2025-03-14 21:05:51 -07:00
Krish Dholakia
82f31beee5
Merge pull request #9271 from BerriAI/litellm_rc_03_14_2025_patch_1
Litellm rc 03 14 2025 patch 1
2025-03-14 20:57:22 -07:00
Krrish Dholakia
b15c06ee94 fix(team_endpoints.py): fix linting error 2025-03-14 20:51:21 -07:00
Krish Dholakia
59fd58643b
Merge pull request #9261 from briandevvn/fix_ollama_pt
Fix "system" role has become unacceptable in ollama
2025-03-14 20:13:28 -07:00
Krrish Dholakia
26226d475f feat(proxy_server.py): support retrieving models for a team, if user is a member - via /models?team_id
Allows user to see team models on UI when creating a key
2025-03-14 19:34:06 -07:00
Ishaan Jaff
6787d0dabe test_model_connection 2025-03-14 18:33:49 -07:00
Krrish Dholakia
621d193727 build: new ui build 2025-03-14 18:18:03 -07:00
Ishaan Jaff
5a6da56058 fix endpoint_data 2025-03-14 17:21:01 -07:00
Krrish Dholakia
a93471a3c7 feat(team_endpoints.py): unfurl 'all-proxy-models' on team info endpoints 2025-03-14 16:54:24 -07:00
Krrish Dholakia
08abee1990 fix: fix linting error 2025-03-14 14:17:28 -07:00
Krrish Dholakia
f089b1e23f feat(endpoints.py): support adding credentials by model id
Allows user to reuse existing model credentials
2025-03-14 12:32:32 -07:00
Ishaan Jaff
d67fc03e20 fix endpoint 2025-03-14 12:26:17 -07:00
Krrish Dholakia
605a4d1121 feat(endpoints.py): enable retrieving existing credentials by model name
Enables reusing existing credentials
2025-03-14 12:02:50 -07:00
Ishaan Jaff
c84717c9e0 add health/test_connection 2025-03-14 11:30:24 -07:00
Ishaan Jaff
6742d3cb10 fix route llm request to allow non-router models 2025-03-14 11:10:10 -07:00
Brian Dev
12db28b0af Support 'system' role ollama 2025-03-15 00:55:18 +07:00
Krrish Dholakia
6629354329 fix(endpoints.py): update credentials should update before storing 2025-03-14 10:42:17 -07:00
Krrish Dholakia
fc0d21794d style: cleanup credential from leftnav - now in models tab 2025-03-14 10:14:21 -07:00
Ishaan Jaff
c36e5cae50 backend instant delete model 2025-03-14 10:12:50 -07:00
Lucas Raschek
56d3e75b33 Map total tokens to prompt_tokens too 2025-03-14 18:04:43 +01:00
Ishaan Jaff
c2d290785a ui new build 2025-03-13 21:44:32 -07:00
Ishaan Jaff
276a7089df
Merge pull request #9220 from BerriAI/litellm_qa_responses_api
[Fixes] Responses API - allow /responses and subpaths as LLM API route + Add exception mapping for responses API
2025-03-13 21:36:59 -07:00
Ishaan Jaff
241a36a74f
Merge pull request #9222 from BerriAI/litellm_snowflake_pr_mar_13
[Feat] Add Snowflake Cortex to LiteLLM
2025-03-13 21:35:39 -07:00
Krish Dholakia
f89dbb8ab3
Merge branch 'main' into litellm_dev_03_13_2025_p3 2025-03-13 20:12:16 -07:00
Ishaan Jaff
69b47cf738 fix code quality check 2025-03-13 20:10:41 -07:00
Ishaan Jaff
7827c275ba exception_type 2025-03-13 20:09:32 -07:00
Krish Dholakia
e8c67f25e3
Merge pull request #9221 from BerriAI/litellm_dev_03_13_2025_p2
Support bedrock converse cache token tracking
2025-03-13 20:08:33 -07:00
Krish Dholakia
fd8a5960ec
Merge pull request #9216 from BerriAI/litellm_dev_03_12_2025_contributor_prs_p2
Litellm dev 03 12 2025 contributor prs p2
2025-03-13 20:03:57 -07:00
Krrish Dholakia
58a7351a73 fix: fix linting errors 2025-03-13 19:40:18 -07:00
Krrish Dholakia
8a6e4715aa feat(converse_transformation.py): fix type for bedrock cache usage block 2025-03-13 19:33:22 -07:00
Krrish Dholakia
f2d0aaacbc fix: fix linting errors 2025-03-13 19:26:46 -07:00