Krrish Dholakia
|
673782ce48
|
test(test_auth_checks.py): add unit tests for customer max budget check
|
2024-05-29 18:03:59 -07:00 |
|
Krish Dholakia
|
8556f4afb0
|
Merge pull request #3902 from BerriAI/litellm_openai_compatible_endpoint_fix
fix(openai.py): only allow 'user' as optional param if openai model
|
2024-05-29 16:34:23 -07:00 |
|
Krish Dholakia
|
f360320ecd
|
Merge pull request #3906 from BerriAI/litellm_customer_webhook_event
feat(proxy_server.py): emit webhook event whenever customer spend is tracked
|
2024-05-29 16:34:16 -07:00 |
|
Krish Dholakia
|
77cc9cded9
|
Revert "fix: Log errors in Traceloop Integration (reverts previous revert)"
|
2024-05-29 16:30:09 -07:00 |
|
Krrish Dholakia
|
1d18ca6a7d
|
fix(router.py): security fix - don't show api key in invalid model setup error message
|
2024-05-29 16:14:57 -07:00 |
|
Krrish Dholakia
|
f729370890
|
feat(proxy_server.py): emit webhook event whenever customer spend is tracked
Closes https://github.com/BerriAI/litellm/issues/3903
|
2024-05-29 15:59:32 -07:00 |
|
Krrish Dholakia
|
3d32b00821
|
fix(openai.py): only allow 'user' as optional param if openai model
|
2024-05-29 15:15:02 -07:00 |
|
Krish Dholakia
|
c76deb8f76
|
Merge pull request #3846 from nirga/revert-3831-revert-3780-traceloop-failures
fix: Log errors in Traceloop Integration (reverts previous revert)
|
2024-05-29 08:54:01 -07:00 |
|
Ishaan Jaff
|
dc8480a5e9
|
test_batch_completion_multiple_models_multiple_messages
|
2024-05-28 22:54:54 -07:00 |
|
Krish Dholakia
|
e838bd1c79
|
Merge branch 'main' into litellm_batch_completions
|
2024-05-28 22:38:05 -07:00 |
|
Ishaan Jaff
|
6e9cc26ca1
|
Merge pull request #3888 from BerriAI/litellm_add_files_proxy
[Feat] LiteLLM Proxy Add `POST /v1/files` and `GET /v1/files`
|
2024-05-28 22:36:33 -07:00 |
|
Ishaan Jaff
|
473ec66b84
|
feat - router add abatch_completion
|
2024-05-28 22:19:33 -07:00 |
|
Krrish Dholakia
|
e3000504f9
|
fix(router.py): support batch completions fastest response streaming
|
2024-05-28 21:51:09 -07:00 |
|
Krrish Dholakia
|
1ebae6e7b0
|
fix(router.py): support comma-separated model list for batch completion fastest response
|
2024-05-28 21:34:37 -07:00 |
|
Ishaan Jaff
|
ca8163bbba
|
feat - add afile_content, file_content
|
2024-05-28 20:58:22 -07:00 |
|
Krrish Dholakia
|
ecd182eb6a
|
feat(router.py): support fastest response batch completion call
returns fastest response. cancels others.
|
2024-05-28 19:44:41 -07:00 |
|
Ishaan Jaff
|
6bf6059b3e
|
Merge pull request #3882 from BerriAI/litellm_add_batches_sdk
[BETA-Feature] Add OpenAI `v1/batches` Support on LiteLLM SDK
|
2024-05-28 19:38:12 -07:00 |
|
Ishaan Jaff
|
5e9d024c2c
|
fix - test openai batches
|
2024-05-28 17:29:13 -07:00 |
|
Krish Dholakia
|
b5c89a249a
|
Merge pull request #3883 from BerriAI/litellm_anthropic_parallel_streaming_fix
fix(anthropic.py): fix parallel streaming on anthropic.py
|
2024-05-28 17:21:57 -07:00 |
|
Ishaan Jaff
|
6688215c18
|
feat - add aretrieve_batch
|
2024-05-28 17:12:41 -07:00 |
|
Ishaan Jaff
|
1ef7cd923c
|
feat - add acreate_batch
|
2024-05-28 17:03:29 -07:00 |
|
Ishaan Jaff
|
758ed9e923
|
feat - add litellm.acreate_file
|
2024-05-28 16:47:27 -07:00 |
|
Krrish Dholakia
|
324bf027f5
|
fix(anthropic.py): fix parallel streaming on anthropic.py
prevent parallel requests from cancelling each other
Fixes https://github.com/BerriAI/litellm/issues/3881
|
2024-05-28 16:29:09 -07:00 |
|
Ishaan Jaff
|
938f4703c3
|
feat- add retrieve_batch()
|
2024-05-28 15:55:32 -07:00 |
|
Ishaan Jaff
|
38285e53c3
|
working create_batch
|
2024-05-28 15:45:23 -07:00 |
|
Ishaan Jaff
|
d5dbf084ed
|
feat - import batches in __init__
|
2024-05-28 15:35:11 -07:00 |
|
Ishaan Jaff
|
0af4c9206f
|
test - openai batches file
|
2024-05-28 14:54:58 -07:00 |
|
frob
|
c6af8977f6
|
Fix typo
|
2024-05-28 17:00:08 +02:00 |
|
frob
|
19194ece78
|
Add tests for ollama image handling
|
2024-05-28 16:48:15 +02:00 |
|
Krrish Dholakia
|
6b50e656b8
|
fix(main.py): pass extra headers through for async calls
|
2024-05-27 19:11:40 -07:00 |
|
Krrish Dholakia
|
24eb79da91
|
test(test_bedrock_completion.py): refactor test bedrock headers test
|
2024-05-27 19:01:07 -07:00 |
|
Krrish Dholakia
|
4408b717f0
|
fix(parallel_request_limiter.py): fix user+team tpm/rpm limit check
Closes https://github.com/BerriAI/litellm/issues/3788
|
2024-05-27 08:48:23 -07:00 |
|
Nir Gazit
|
7602c6f436
|
Revert "Revert "Log errors in Traceloop Integration""
|
2024-05-26 12:01:10 +03:00 |
|
Krrish Dholakia
|
dde6af7420
|
test(test_alerting.py): fix test
|
2024-05-25 21:48:01 -07:00 |
|
Krish Dholakia
|
4e230f0a2e
|
Merge pull request #3844 from BerriAI/litellm_region_based_alerts
feat(slack_alerting.py): enable provider-region based alerting
|
2024-05-25 21:03:16 -07:00 |
|
Ishaan Jaff
|
7cc63a77c6
|
fix - router debug logs
|
2024-05-25 17:23:37 -07:00 |
|
Krrish Dholakia
|
c2f19d631e
|
feat(slack_alerting.py): enable provider-region based alerting
|
2024-05-25 16:55:46 -07:00 |
|
Ishaan Jaff
|
59cb306d47
|
fix - __team_info
|
2024-05-25 15:36:20 -07:00 |
|
Ishaan Jaff
|
e2d3c0c846
|
fix - updating team
|
2024-05-25 15:31:25 -07:00 |
|
Ishaan Jaff
|
22d8eaab93
|
test - new_team
|
2024-05-25 15:20:49 -07:00 |
|
Krrish Dholakia
|
d2e14ca833
|
fix(bedrock_httpx.py): fix bedrock ptu model id str encoding
Fixes https://github.com/BerriAI/litellm/issues/3805
|
2024-05-25 10:54:01 -07:00 |
|
Krrish Dholakia
|
00af8e350f
|
fix(bedrock_httpx.py): support bedrock ptu's
Fixes https://github.com/BerriAI/litellm/issues/3805
|
2024-05-24 23:02:04 -07:00 |
|
Krrish Dholakia
|
281906ff33
|
fix(vertex_ai.py): handle nested content dictionary for assistant message
openai allows the assistant content message to also be a list of dictionaries, handle that
|
2024-05-24 22:47:47 -07:00 |
|
Ishaan Jaff
|
0083776a14
|
Revert "Log errors in Traceloop Integration"
|
2024-05-24 21:25:17 -07:00 |
|
Ishaan Jaff
|
b2feb9a8ec
|
ci / cd - run again
|
2024-05-24 21:13:50 -07:00 |
|
Ishaan Jaff
|
2e5c625f8b
|
ci/cd run again
|
2024-05-24 20:56:52 -07:00 |
|
Ishaan Jaff
|
9d3b2e7e0d
|
test alerting
|
2024-05-24 20:54:17 -07:00 |
|
Krish Dholakia
|
d25ed9c4d3
|
Merge pull request #3828 from BerriAI/litellm_outage_alerting
fix(slack_alerting.py): support region based outage alerting
|
2024-05-24 19:13:17 -07:00 |
|
Krrish Dholakia
|
8dec87425e
|
feat(slack_alerting.py): refactor region outage alerting to do model based alerting instead
Unable to extract azure region from api base, makes sense to start with model alerting and then move to region
|
2024-05-24 19:10:33 -07:00 |
|
Krrish Dholakia
|
2cdb0584d1
|
fix(slack_alerting.py): fixes for outage alerting
|
2024-05-24 17:17:17 -07:00 |
|