Commit graph

3040 commits

Author SHA1 Message Date
Krrish Dholakia
673782ce48 test(test_auth_checks.py): add unit tests for customer max budget check 2024-05-29 18:03:59 -07:00
Krish Dholakia
8556f4afb0
Merge pull request #3902 from BerriAI/litellm_openai_compatible_endpoint_fix
fix(openai.py): only allow 'user' as optional param if openai model
2024-05-29 16:34:23 -07:00
Krish Dholakia
f360320ecd
Merge pull request #3906 from BerriAI/litellm_customer_webhook_event
feat(proxy_server.py): emit webhook event whenever customer spend is tracked
2024-05-29 16:34:16 -07:00
Krish Dholakia
77cc9cded9
Revert "fix: Log errors in Traceloop Integration (reverts previous revert)" 2024-05-29 16:30:09 -07:00
Krrish Dholakia
1d18ca6a7d fix(router.py): security fix - don't show api key in invalid model setup error message 2024-05-29 16:14:57 -07:00
Krrish Dholakia
f729370890 feat(proxy_server.py): emit webhook event whenever customer spend is tracked
Closes https://github.com/BerriAI/litellm/issues/3903
2024-05-29 15:59:32 -07:00
Krrish Dholakia
3d32b00821 fix(openai.py): only allow 'user' as optional param if openai model 2024-05-29 15:15:02 -07:00
Krish Dholakia
c76deb8f76
Merge pull request #3846 from nirga/revert-3831-revert-3780-traceloop-failures
fix: Log errors in Traceloop Integration (reverts previous revert)
2024-05-29 08:54:01 -07:00
Ishaan Jaff
dc8480a5e9 test_batch_completion_multiple_models_multiple_messages 2024-05-28 22:54:54 -07:00
Krish Dholakia
e838bd1c79
Merge branch 'main' into litellm_batch_completions 2024-05-28 22:38:05 -07:00
Ishaan Jaff
6e9cc26ca1
Merge pull request #3888 from BerriAI/litellm_add_files_proxy
[Feat] LiteLLM Proxy Add `POST /v1/files` and `GET /v1/files`
2024-05-28 22:36:33 -07:00
Ishaan Jaff
473ec66b84 feat - router add abatch_completion 2024-05-28 22:19:33 -07:00
Krrish Dholakia
e3000504f9 fix(router.py): support batch completions fastest response streaming 2024-05-28 21:51:09 -07:00
Krrish Dholakia
1ebae6e7b0 fix(router.py): support comma-separated model list for batch completion fastest response 2024-05-28 21:34:37 -07:00
Ishaan Jaff
ca8163bbba feat - add afile_content, file_content 2024-05-28 20:58:22 -07:00
Krrish Dholakia
ecd182eb6a feat(router.py): support fastest response batch completion call
returns fastest response. cancels others.
2024-05-28 19:44:41 -07:00
Ishaan Jaff
6bf6059b3e
Merge pull request #3882 from BerriAI/litellm_add_batches_sdk
[BETA-Feature] Add OpenAI `v1/batches` Support on LiteLLM SDK
2024-05-28 19:38:12 -07:00
Ishaan Jaff
5e9d024c2c fix - test openai batches 2024-05-28 17:29:13 -07:00
Krish Dholakia
b5c89a249a
Merge pull request #3883 from BerriAI/litellm_anthropic_parallel_streaming_fix
fix(anthropic.py): fix parallel streaming on anthropic.py
2024-05-28 17:21:57 -07:00
Ishaan Jaff
6688215c18 feat - add aretrieve_batch 2024-05-28 17:12:41 -07:00
Ishaan Jaff
1ef7cd923c feat - add acreate_batch 2024-05-28 17:03:29 -07:00
Ishaan Jaff
758ed9e923 feat - add litellm.acreate_file 2024-05-28 16:47:27 -07:00
Krrish Dholakia
324bf027f5 fix(anthropic.py): fix parallel streaming on anthropic.py
prevent parallel requests from cancelling each other

Fixes https://github.com/BerriAI/litellm/issues/3881
2024-05-28 16:29:09 -07:00
Ishaan Jaff
938f4703c3 feat- add retrieve_batch() 2024-05-28 15:55:32 -07:00
Ishaan Jaff
38285e53c3 working create_batch 2024-05-28 15:45:23 -07:00
Ishaan Jaff
d5dbf084ed feat - import batches in __init__ 2024-05-28 15:35:11 -07:00
Ishaan Jaff
0af4c9206f test - openai batches file 2024-05-28 14:54:58 -07:00
frob
c6af8977f6
Fix typo 2024-05-28 17:00:08 +02:00
frob
19194ece78
Add tests for ollama image handling 2024-05-28 16:48:15 +02:00
Krrish Dholakia
6b50e656b8 fix(main.py): pass extra headers through for async calls 2024-05-27 19:11:40 -07:00
Krrish Dholakia
24eb79da91 test(test_bedrock_completion.py): refactor test bedrock headers test 2024-05-27 19:01:07 -07:00
Krrish Dholakia
4408b717f0 fix(parallel_request_limiter.py): fix user+team tpm/rpm limit check
Closes https://github.com/BerriAI/litellm/issues/3788
2024-05-27 08:48:23 -07:00
Nir Gazit
7602c6f436 Revert "Revert "Log errors in Traceloop Integration"" 2024-05-26 12:01:10 +03:00
Krrish Dholakia
dde6af7420 test(test_alerting.py): fix test 2024-05-25 21:48:01 -07:00
Krish Dholakia
4e230f0a2e
Merge pull request #3844 from BerriAI/litellm_region_based_alerts
feat(slack_alerting.py): enable provider-region based alerting
2024-05-25 21:03:16 -07:00
Ishaan Jaff
7cc63a77c6 fix - router debug logs 2024-05-25 17:23:37 -07:00
Krrish Dholakia
c2f19d631e feat(slack_alerting.py): enable provider-region based alerting 2024-05-25 16:55:46 -07:00
Ishaan Jaff
59cb306d47 fix - __team_info 2024-05-25 15:36:20 -07:00
Ishaan Jaff
e2d3c0c846 fix - updating team 2024-05-25 15:31:25 -07:00
Ishaan Jaff
22d8eaab93 test - new_team 2024-05-25 15:20:49 -07:00
Krrish Dholakia
d2e14ca833 fix(bedrock_httpx.py): fix bedrock ptu model id str encoding
Fixes https://github.com/BerriAI/litellm/issues/3805
2024-05-25 10:54:01 -07:00
Krrish Dholakia
00af8e350f fix(bedrock_httpx.py): support bedrock ptu's
Fixes https://github.com/BerriAI/litellm/issues/3805
2024-05-24 23:02:04 -07:00
Krrish Dholakia
281906ff33 fix(vertex_ai.py): handle nested content dictionary for assistant message
openai allows the assistant content message to also be a list of dictionaries, handle that
2024-05-24 22:47:47 -07:00
Ishaan Jaff
0083776a14
Revert "Log errors in Traceloop Integration" 2024-05-24 21:25:17 -07:00
Ishaan Jaff
b2feb9a8ec ci / cd - run again 2024-05-24 21:13:50 -07:00
Ishaan Jaff
2e5c625f8b ci/cd run again 2024-05-24 20:56:52 -07:00
Ishaan Jaff
9d3b2e7e0d test alerting 2024-05-24 20:54:17 -07:00
Krish Dholakia
d25ed9c4d3
Merge pull request #3828 from BerriAI/litellm_outage_alerting
fix(slack_alerting.py): support region based outage alerting
2024-05-24 19:13:17 -07:00
Krrish Dholakia
8dec87425e feat(slack_alerting.py): refactor region outage alerting to do model based alerting instead
Unable to extract azure region from api base, makes sense to start with model alerting and then move to region
2024-05-24 19:10:33 -07:00
Krrish Dholakia
2cdb0584d1 fix(slack_alerting.py): fixes for outage alerting 2024-05-24 17:17:17 -07:00