Commit graph

1124 commits

Author SHA1 Message Date
Krrish Dholakia
0f1c40d698 fix(main.py): trigger new build 2024-06-11 15:13:50 -07:00
wslee
61badd8fd6 change friendli_ai -> friendliai 2024-06-11 16:17:30 +09:00
wslee
99e2050125 resolve comments 2024-06-11 14:49:39 +09:00
wslee
2ba421fbe0 add friendli_ai provider 2024-06-10 17:27:15 +09:00
Krrish Dholakia
39ee6be477 fix(utils.py): improved predibase exception mapping
adds unit testing + better coverage for predibase errors
2024-06-08 14:32:43 -07:00
Ishaan Jaff
1cbd36433b fix mock_completion 2024-06-07 19:10:05 -07:00
Ishaan Jaff
92841dfe1b Merge branch 'main' into litellm_security_fix 2024-06-07 16:52:25 -07:00
Krrish Dholakia
de98bd939c fix(test_custom_callbacks_input.py): unit tests for 'turn_off_message_logging'
ensure no raw request is logged either
2024-06-07 15:39:15 -07:00
Ishaan Jaff
80def35a04 Merge pull request #4065 from BerriAI/litellm_use_common_func
[Refactor] - Refactor proxy_server.py to use common function for `add_litellm_data_to_request`
2024-06-07 14:02:17 -07:00
Ishaan Jaff
860c9b52b6 Merge branch 'main' into litellm_svc_logger 2024-06-07 14:01:54 -07:00
Ishaan Jaff
8106a6dc9b fix simplify - pass litellm_parent_otel_span 2024-06-07 13:48:21 -07:00
Ishaan Jaff
0f99d47d87 use litellm_parent_otel_span as litellm_param 2024-06-07 08:54:28 -07:00
Krish Dholakia
7bf5c61007 Merge branch 'main' into litellm_bedrock_converse_api 2024-06-07 08:49:52 -07:00
Krrish Dholakia
12ed3dc911 refactor(main.py): only route anthropic calls through converse api
v0 scope let's move function calling to converse api
2024-06-07 08:47:51 -07:00
Krrish Dholakia
bad5cde7c5 test(main.py): test cicd 2024-06-07 08:18:37 -07:00
Krrish Dholakia
3f4b617767 style(main.py): trigger new build 2024-06-07 08:10:28 -07:00
Krrish Dholakia
f8b5aa3df6 fix(bedrock_httpx.py): working claude 3 function calling 2024-06-06 20:12:41 -07:00
Krrish Dholakia
e391e30285 refactor: replace 'traceback.print_exc()' with logging library
allows error logs to be in json format for otel logging
2024-06-06 13:47:43 -07:00
Krish Dholakia
e678dce88b Merge pull request #4009 from BerriAI/litellm_fix_streaming_cost_cal
fix(utils.py): fix cost calculation for openai-compatible streaming object
2024-06-04 21:00:22 -07:00
Krrish Dholakia
11b44192c2 fix(main.py): check verify ssl on custom endpoint call 2024-06-04 17:12:42 -07:00
Krrish Dholakia
7432c6a4d9 fix(utils.py): fix cost calculation for openai-compatible streaming object 2024-06-04 10:36:25 -07:00
Krrish Dholakia
661b67e71c fix(main.py): fix typing for image gen response 2024-06-04 08:29:30 -07:00
Krrish Dholakia
46a0ac7953 fix(main.py): fix ahealth_check to infer mode when custom_llm_provider/model_name used 2024-06-03 14:06:36 -07:00
Krrish Dholakia
594daef07a fix(utils.py): correctly instrument passing through api version in optional param check 2024-06-01 19:31:52 -07:00
Krrish Dholakia
5d3a0ace4b fix(openai.py): fix client caching logic 2024-06-01 16:45:56 -07:00
Krrish Dholakia
69244aabf3 fix(http_handler.py): allow setting ca bundle path 2024-06-01 14:48:53 -07:00
Krrish Dholakia
de62c5f565 docs(assistants.md): add assistants api to docs 2024-06-01 10:30:07 -07:00
Krish Dholakia
1529f665cc Merge pull request #3954 from BerriAI/litellm_simple_request_prioritization
feat(scheduler.py): add request prioritization scheduler
2024-05-31 23:29:09 -07:00
Krrish Dholakia
6221fabecf fix(router.py): fix cooldown logic for usage-based-routing-v2 pre-call-checks 2024-05-31 21:32:01 -07:00
Krrish Dholakia
3896e3e88f fix: fix streaming with httpx client
prevent overwriting streams in parallel streaming calls
2024-05-31 10:55:18 -07:00
Krrish Dholakia
6b4153ff03 fix(main.py): add logging to audio_transcription calls 2024-05-30 16:57:11 -07:00
Krrish Dholakia
eb159b64e1 fix(openai.py): fix openai response for /audio/speech endpoint 2024-05-30 16:41:06 -07:00
Krrish Dholakia
1e89a1f56e feat(main.py): support openai tts endpoint
Closes https://github.com/BerriAI/litellm/issues/3094
2024-05-30 14:28:28 -07:00
Krrish Dholakia
5f01dce284 fix(main.py): pass api key and api base to openai.py for audio transcription call 2024-05-29 21:29:01 -07:00
Giri Tatavarty
2d8b4928bf #Fixed mypy errors. The requests package and stubs need to be imported - waiting to hear from Ishaan/Krrish before changing requirements.txt 2024-05-29 15:08:56 -07:00
Ishaan Jaff
93bf4c2dc4 Revert "Added support for Triton chat completion using trtlllm generate endpo…" 2024-05-29 13:42:49 -07:00
Ishaan Jaff
64d050cadd Merge pull request #3895 from giritatavarty-8451/litellm_triton_chatcompletion_support
Added support for Triton chat completion using trtlllm generate endpo…
2024-05-29 12:50:31 -07:00
Krrish Dholakia
00cce4ea95 build(config.yml): add pillow to ci/cd 2024-05-28 21:39:09 -07:00
Krrish Dholakia
792b25c772 feat(proxy_server.py): enable batch completion fastest response calls on proxy
introduces new `fastest_response` flag for enabling the call
2024-05-28 20:09:31 -07:00
Giri Tatavarty
ff18d93a3a Added support for Triton chat completion using trtlllm generate endpoint and custom infer endpoint 2024-05-28 07:54:11 -07:00
Krrish Dholakia
4aa7e0b17c fix(main.py): pass extra headers through for async calls 2024-05-27 19:11:40 -07:00
Krrish Dholakia
68a8b23b59 fix(bedrock_httpx.py): fix bedrock ptu model id str encoding
Fixes https://github.com/BerriAI/litellm/issues/3805
2024-05-25 10:54:01 -07:00
Krish Dholakia
40791ee1f8 Merge pull request #3828 from BerriAI/litellm_outage_alerting
fix(slack_alerting.py): support region based outage alerting
2024-05-24 19:13:17 -07:00
Krrish Dholakia
4536ed6f6e feat(slack_alerting.py): refactor region outage alerting to do model based alerting instead
Unable to extract azure region from api base, makes sense to start with model alerting and then move to region
2024-05-24 19:10:33 -07:00
Krrish Dholakia
7368406c24 fix(slack_alerting.py): support region based outage alerting 2024-05-24 16:59:16 -07:00
Ishaan Jaff
a731e00c6e Merge pull request #3462 from ffreemt/main
Add return_exceptions to batch_completion (retry)
2024-05-24 09:19:10 -07:00
ffreemt
ae6834e97a Make return-exceptions as default behavior in litellm.batch_completion 2024-05-24 11:09:11 +08:00
Krrish Dholakia
e3c5e004c5 feat(databricks.py): add embedding model support 2024-05-23 18:22:03 -07:00
Krrish Dholakia
143a44823a feat(databricks.py): adds databricks support - completion, async, streaming
Closes https://github.com/BerriAI/litellm/issues/2160
2024-05-23 16:29:46 -07:00
Krrish Dholakia
4795c56f84 feat(anthropic.py): support anthropic 'tool_choice' param
Closes https://github.com/BerriAI/litellm/issues/3752
2024-05-21 17:50:44 -07:00