Commit graph

1610 commits

Author SHA1 Message Date
Krrish Dholakia
00b338cb9c test: refactor testing to handle routing correctly 2025-03-18 12:24:12 -07:00
Krrish Dholakia
267084a1af test(test_get_llm_provider.py): cover scenario where xai not in model name 2025-03-18 11:04:59 -07:00
Krrish Dholakia
aeec703c4e test(test_get_llm_provider.py): Minimal repro for https://github.com/BerriAI/litellm/issues/9291 2025-03-18 10:35:50 -07:00
Ishaan Jaff
f2026ef907 fix - correctly re-use azure openai client 2025-03-18 09:51:28 -07:00
Nate Mar
9f9e555342 Update test_arize_ai.py 2025-03-18 04:04:44 -07:00
Nate Mar
6b821fbbc0 restore previous formatting 2025-03-18 04:02:43 -07:00
Nate Mar
a1d188ba5e Fix test and add comments 2025-03-18 03:46:53 -07:00
Nate Mar
434e262b8c revert space_key change and add tests for arize integration 2025-03-18 01:40:10 -07:00
Krrish Dholakia
b341b788a5 test: update testing 2025-03-17 22:34:08 -07:00
Krish Dholakia
bcbb88d802
Merge pull request #9326 from andjsmi/main
Modify completion handler for SageMaker to use payload from `prepared_request`
2025-03-17 22:16:43 -07:00
Ishaan Jaff
d42f84dd4e
Merge pull request #9328 from BerriAI/litellm_control_internal_user_settings_ui
[UI] - Allow controlling default internal user settings on ui
2025-03-17 22:07:47 -07:00
Krish Dholakia
cd5024f3b1
Merge pull request #9333 from BerriAI/litellm_dev_03_17_2025_p2
fix(ollama/completions/transformation.py): pass prompt, untemplated o…
2025-03-17 21:48:30 -07:00
Ishaan Jaff
5400615ce8
Merge pull request #9329 from BerriAI/litellm_fix_reset_budget_job
[Bug fix] Reset Budget Job
2025-03-17 21:46:08 -07:00
Nate Mar
35e0856f11 Fix wrong import and use space_id instead of space_key for Arize integration 2025-03-17 20:37:28 -07:00
Krish Dholakia
27768a4866
Merge pull request #9330 from BerriAI/litellm_dev_03_17_2025_p1
Litellm dev 03 17 2025 p1
2025-03-17 19:57:25 -07:00
Ishaan Jaff
fbace8d041 unit test reset budget job 2025-03-17 19:54:51 -07:00
Ishaan Jaff
63a32ff02c test_reset_budget_job 2025-03-17 19:37:09 -07:00
Krrish Dholakia
22faf7d232 fix(ollama/completions/transformation.py): pass prompt, untemplated on /completions request
Fixes https://github.com/BerriAI/litellm/issues/6900
2025-03-17 18:35:44 -07:00
Ishaan Jaff
b4568762d2 update_internal_user_settings 2025-03-17 18:26:42 -07:00
Krrish Dholakia
359b8298f8 test(test_assemble_streaming_responses.py): update test to use correct type 2025-03-17 18:17:01 -07:00
Krrish Dholakia
c4b2e0ae3d fix(streaming_handler.py): support logging complete streaming response on cache hit 2025-03-17 18:10:39 -07:00
Krrish Dholakia
dd9e79adbd fix(streaming_handler.py): emit deep copy of completed chunk 2025-03-17 17:26:21 -07:00
Andrew Smith
1b2439cb82
Merge branch 'BerriAI:main' into main 2025-03-18 11:10:02 +11:00
Andrew Smith
d5274efe62
Update test_sagemaker.py to match changed parameters 2025-03-18 11:09:52 +11:00
Krrish Dholakia
8e27b2026a fix(http_handler.py): support reading ssl security level from env var
Allows user to specify lower security settings
2025-03-17 15:48:31 -07:00
Krrish Dholakia
078e2d341b feat(cost_calculator.py): support reading litellm response cost header in client sdk
allows consistent cost tracking when sdk is calling proxy
2025-03-17 15:12:01 -07:00
Krrish Dholakia
db92956ae3 fix(redis_cache.py): add 5s default timeout 2025-03-17 14:27:36 -07:00
Krish Dholakia
d0d8ec2c40
Merge branch 'main' into litellm_dev_03_16_2025_p1 2025-03-17 10:02:53 -07:00
Krrish Dholakia
8618295911 test: loosen test 2025-03-17 09:44:22 -07:00
Krrish Dholakia
d01361747d test: make test less flaky 2025-03-17 09:00:15 -07:00
Krrish Dholakia
a5b497667c fix(logging_utils.py): revert change 2025-03-16 21:04:41 -07:00
Krrish Dholakia
c0a76427d2 fix(streaming_handler.py): pass complete streaming response on completion 2025-03-16 20:22:12 -07:00
Krrish Dholakia
937f0e2ca2 test(test_bedrock_completion.py): ensure model id in model name just works 2025-03-15 14:09:37 -07:00
Krrish Dholakia
5dc46f0cf7 fix(converse_transformation.py): fix encoding model 2025-03-15 14:03:37 -07:00
Krrish Dholakia
612d5a284d refactor(litellm_logging.py): delegate returning a complete response to the streaming_handler
Removes incorrect logic for calculating complete streaming response from litellm logging
2025-03-15 09:55:33 -07:00
Krrish Dholakia
cc82d42d25 test(test_streaming_handler.py): add unit test to ensure model response stream with usage is always used 2025-03-15 09:35:48 -07:00
Krrish Dholakia
82252ecac1 test(test_streaming_handler.py): assert chunk is non-empty when usage block given 2025-03-15 09:26:01 -07:00
Krrish Dholakia
dd2c980d5b fix(utils.py): Prevents final chunk w/ usage from being ignored
Fixes https://github.com/BerriAI/litellm/issues/7112
2025-03-15 09:12:14 -07:00
Krish Dholakia
d4caaae1be
Merge pull request #9274 from BerriAI/litellm_contributor_rebase_branch
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 43s
Helm unit test / unit-test (push) Successful in 50s
Litellm contributor rebase branch
2025-03-14 21:57:49 -07:00
Krish Dholakia
82f31beee5
Merge pull request #9271 from BerriAI/litellm_rc_03_14_2025_patch_1
Litellm rc 03 14 2025 patch 1
2025-03-14 20:57:22 -07:00
Krrish Dholakia
6ed995952f fix: fix test 2025-03-14 20:28:50 -07:00
Krish Dholakia
59fd58643b
Merge pull request #9261 from briandevvn/fix_ollama_pt
Fix "system" role has become unacceptable in ollama
2025-03-14 20:13:28 -07:00
Krrish Dholakia
c2f01b0fdc fix(router.py): add new test 2025-03-14 14:23:45 -07:00
Brian Dev
ff3f79e468 Update test_ollama_pt 2025-03-15 01:07:15 +07:00
Brian Dev
51589364a1 Add new line 2025-03-15 01:03:12 +07:00
Brian Dev
12db28b0af Support 'system' role ollama 2025-03-15 00:55:18 +07:00
Lucas Raschek
88798d4b3d Add unit test 2025-03-14 18:14:41 +01:00
Ishaan Jaff
276a7089df
Merge pull request #9220 from BerriAI/litellm_qa_responses_api
[Fixes] Responses API - allow /responses and subpaths as LLM API route + Add exception mapping for responses API
2025-03-13 21:36:59 -07:00
Ishaan Jaff
241a36a74f
Merge pull request #9222 from BerriAI/litellm_snowflake_pr_mar_13
[Feat] Add Snowflake Cortex to LiteLLM
2025-03-13 21:35:39 -07:00
Krish Dholakia
f89dbb8ab3
Merge branch 'main' into litellm_dev_03_13_2025_p3 2025-03-13 20:12:16 -07:00