Commit graph

218 commits

Author SHA1 Message Date
Krrish Dholakia
12ed3dc911 refactor(main.py): only route anthropic calls through converse api
v0 scope let's move function calling to converse api
2024-06-07 08:47:51 -07:00
Krrish Dholakia
c41b60f6bf feat(bedrock_httpx.py): working bedrock converse api streaming 2024-06-06 22:13:21 -07:00
Krish Dholakia
e678dce88b Merge pull request #4009 from BerriAI/litellm_fix_streaming_cost_cal
fix(utils.py): fix cost calculation for openai-compatible streaming object
2024-06-04 21:00:22 -07:00
Krrish Dholakia
1336957077 fix(utils.py): fix 2024-06-04 19:41:20 -07:00
Krrish Dholakia
e279498970 feat(utils.py): support 'stream_options' param across all providers
Closes https://github.com/BerriAI/litellm/issues/3553
2024-06-04 19:03:26 -07:00
Krrish Dholakia
d74ccc6c84 fix(utils.py): fix stream options to return consistent response object 2024-06-04 18:17:45 -07:00
Krrish Dholakia
7432c6a4d9 fix(utils.py): fix cost calculation for openai-compatible streaming object 2024-06-04 10:36:25 -07:00
Ishaan Jaff
e21408717c test - test gpt-3.5-turbo on test_parallel_streaming_requests 2024-06-01 20:56:33 -07:00
Krrish Dholakia
594daef07a fix(utils.py): correctly instrument passing through api version in optional param check 2024-06-01 19:31:52 -07:00
Krrish Dholakia
ca82b82d4b fix(test_scheduler.py): fix test 2024-06-01 11:30:26 -07:00
Krrish Dholakia
3896e3e88f fix: fix streaming with httpx client
prevent overwriting streams in parallel streaming calls
2024-05-31 10:55:18 -07:00
Krrish Dholakia
ccee9e4eb1 fix(anthropic.py): fix parallel streaming on anthropic.py
prevent parallel requests from cancelling each other

Fixes https://github.com/BerriAI/litellm/issues/3881
2024-05-28 16:29:09 -07:00
Andres Barbaro
abb8e17d4b Update test 2024-05-24 11:48:35 -05:00
Andres Barbaro
33a047315e Improve test 2024-05-24 11:11:43 -05:00
Andres Barbaro
2cce5b3959 Remove unnecessary import 2024-05-24 11:00:29 -05:00
Andres Barbaro
c893674180 Add test to validate chunks with no delta are processed correctly 2024-05-24 10:57:35 -05:00
Krrish Dholakia
143a44823a feat(databricks.py): adds databricks support - completion, async, streaming
Closes https://github.com/BerriAI/litellm/issues/2160
2024-05-23 16:29:46 -07:00
Krrish Dholakia
808bea03db test(test_streaming.py): retry if openai is inconsistent with stream options 2024-05-21 18:51:24 -07:00
Krrish Dholakia
56084d5ac1 fix(bedrock_httpx.py): support mapping for bedrock cohere command r text 2024-05-17 16:13:49 -07:00
Krrish Dholakia
13e4196e3e fix(bedrock_httpx.py): add async support for bedrock amazon, meta, mistral models 2024-05-16 22:39:25 -07:00
Krrish Dholakia
8409b39f0d fix(bedrock_httpx.py): move bedrock ai21 calls to being async 2024-05-16 22:21:30 -07:00
Krrish Dholakia
118fc4ffac fix(bedrock_httpx.py): move anthropic bedrock calls to httpx
Fixing https://github.com/BerriAI/litellm/issues/2921
2024-05-16 21:51:55 -07:00
Krrish Dholakia
e41897808d fix(replicate.py): move replicate calls to being completely async
Closes https://github.com/BerriAI/litellm/issues/3128
2024-05-16 17:24:08 -07:00
Krrish Dholakia
87a21115c5 test: handle watsonx rate limit error 2024-05-13 18:27:39 -07:00
Krish Dholakia
784ae85ba0 Merge branch 'main' into litellm_bedrock_command_r_support 2024-05-11 21:24:42 -07:00
Krrish Dholakia
df34c64464 test(test_streaming.py): fix test 2024-05-11 19:44:47 -07:00
Krrish Dholakia
68596ced04 feat(bedrock_httpx.py): working bedrock command-r sync+async streaming 2024-05-11 19:39:51 -07:00
Krish Dholakia
8ab9c861c9 Merge pull request #3369 from mogith-pn/main
Clarifai-LiteLLM : Added clarifai as LLM Provider.
2024-05-11 09:31:46 -07:00
Krish Dholakia
ddf09a3193 Merge pull request #3552 from BerriAI/litellm_predibase_support
feat(predibase.py): add support for predibase provider
2024-05-09 22:21:16 -07:00
Krrish Dholakia
7c0ab40fd5 feat(predibase.py): support async_completion + streaming (sync + async)
finishes up pr
2024-05-09 17:41:27 -07:00
Ishaan Jaff
27de9e92cc test - stream_options on OpenAI text_completion 2024-05-09 08:41:31 -07:00
Ishaan Jaff
23dc15b9ec test openai stream_options 2024-05-08 21:52:39 -07:00
mogith-pn
c6f9cb9346 Clarifai - Added streaming and async completion support 2024-05-03 14:03:38 +00:00
Krrish Dholakia
a241cdb818 fix(utils.py): fix default values 2024-05-01 22:24:37 -07:00
Krish Dholakia
b7beab2e39 Merge pull request #3270 from simonsanvil/feature/watsonx-integration
(feat) add IBM watsonx.ai as an llm provider
2024-04-27 05:48:34 -07:00
Krrish Dholakia
85a331024a test(test_streaming.py): fix test 2024-04-25 20:22:18 -07:00
Krrish Dholakia
f7d40260fb test: rename test 2024-04-25 20:07:40 -07:00
Krrish Dholakia
2c2e4319ab fix(utils.py): add more logging to identify ci/cd issue 2024-04-25 19:57:24 -07:00
Krrish Dholakia
157dd819f6 fix(utils.py): fix the response object returned when n>1 for stream=true
Fixes https://github.com/BerriAI/litellm/issues/3276
2024-04-25 13:27:29 -07:00
Krrish Dholakia
b10f03706d fix(utils.py): fix streaming to not return usage dict
Fixes https://github.com/BerriAI/litellm/issues/3237
2024-04-24 08:06:07 -07:00
Simon Sanchez Viloria
2dd79eb30e (test) Added completion and embedding tests for watsonx provider 2024-04-24 12:52:29 +02:00
Krish Dholakia
ef471480b2 Merge pull request #3124 from elisalimli/bugfix/add-missing-tool-calls-mistral-messages
Add missing tool_calls and name to messages
2024-04-23 17:25:12 -07:00
alisalim17
25e594dc44 test: add test forfunction calling with mistral large latest to test_streaming.py 2024-04-21 11:27:42 +04:00
Krrish Dholakia
64fe5b146c fix(utils.py): fix azure streaming logic 2024-04-18 07:08:36 -07:00
Krrish Dholakia
caa46ca905 fix(utils.py): fix streaming special character flushing logic 2024-04-17 18:03:40 -07:00
Krrish Dholakia
1b4462ee70 fix(utils.py): ensure streaming output parsing only applied for hf / sagemaker models
selectively applies the <s>
</s> checking
2024-04-17 17:43:41 -07:00
Ishaan Jaff
b00ad976af fix function calling prompt - ask llm to respond in fahrenheit 2024-04-16 21:09:53 -07:00
Krrish Dholakia
1cd0551a1e fix(anthropic_text.py): add support for async text completion calls 2024-04-15 08:15:00 -07:00
Ishaan Jaff
d1d3d932ca Merge pull request #2879 from BerriAI/litellm_async_anthropic_api
[Feat] Async Anthropic API 97.5% lower median latency
2024-04-07 09:56:52 -07:00
Ishaan Jaff
9b8555e482 undo adding next-asyncio 2024-04-06 19:28:51 -07:00