Commit graph

1358 commits

Author SHA1 Message Date
Krrish Dholakia
6a4001c4f4 fix(vertex_ai_llama3.py): Fix llama3 streaming issue
Closes https://github.com/BerriAI/litellm/issues/4885
2024-07-25 22:30:55 -07:00
Krish Dholakia
a5cea7929d Merge branch 'main' into bedrock-llama3.1-405b 2024-07-25 19:29:10 -07:00
Ishaan Jaff
3eea3f0e2d Merge pull request #4884 from Manouchehri/add-mistral-large-2407-bedrock-1
Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock
2024-07-25 19:22:46 -07:00
Krish Dholakia
ff6d153d58 Merge pull request #4879 from fracapuano/main
Add Single-Token predictions support for Replicate
2024-07-25 19:10:57 -07:00
Krish Dholakia
9a42d312b5 Merge pull request #4887 from BerriAI/litellm_custom_llm
feat(custom_llm.py): Support Custom LLM Handlers
2024-07-25 19:05:29 -07:00
Krrish Dholakia
a2de16582a fix(custom_llm.py): pass input params to custom llm 2024-07-25 19:03:52 -07:00
Ishaan Jaff
0b10d40674 feat support audio health checks for azure 2024-07-25 17:30:15 -07:00
Ishaan Jaff
591857082b feat - support health check audio_speech 2024-07-25 17:26:14 -07:00
Ishaan Jaff
56b55b6ab4 fix whisper health check with litellm 2024-07-25 17:22:57 -07:00
Krrish Dholakia
9b1c7066b7 feat(utils.py): support async streaming for custom llm provider 2024-07-25 17:11:57 -07:00
Krrish Dholakia
bf23aac11d feat(utils.py): support sync streaming for custom llm provider 2024-07-25 16:47:32 -07:00
Krrish Dholakia
fe503386ab fix(custom_llm.py): support async completion calls 2024-07-25 15:51:39 -07:00
Krrish Dholakia
54e1ca29b7 feat(custom_llm.py): initial working commit for writing your own custom LLM handler
Fixes https://github.com/BerriAI/litellm/issues/4675

 Also Addresses https://github.com/BerriAI/litellm/discussions/4677
2024-07-25 15:33:05 -07:00
David Manouchehri
213227a7e0 Support tool calling for Llama 3.1 on Amazon bedrock. 2024-07-25 20:36:25 +00:00
David Manouchehri
c7fd02faf3 Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock. 2024-07-25 20:04:03 +00:00
David Manouchehri
97d84ce860 Add Llama 3.1 405b for Bedrock 2024-07-25 19:30:13 +00:00
fracapuano
42ead92211 fix: now supports single tokens prediction 2024-07-25 19:06:07 +02:00
Krrish Dholakia
b94b9dcdea fix(bedrock_httpx.py): fix async client check 2024-07-24 12:43:52 -07:00
David Manouchehri
3ee6baabf7 Add Llama 3.1 for Bedrock. 2024-07-24 05:48:23 +00:00
Krish Dholakia
29eead1add Merge pull request #4845 from BerriAI/litellm_vertex_ai_llama3_1_api
feat(vertex_ai_llama.py): vertex ai llama3.1 api support
2024-07-23 21:51:46 -07:00
Krrish Dholakia
a7779796ef fix(anthropic.py): support openai system message being a list 2024-07-23 21:45:56 -07:00
Krrish Dholakia
23a3be184b build(model_prices_and_context_window.json): add model pricing for vertex ai llama 3.1 api 2024-07-23 17:36:07 -07:00
Krrish Dholakia
778afcee31 feat(vertex_ai_llama.py): vertex ai llama3.1 api support
Initial working commit for vertex ai llama 3.1 api support
2024-07-23 17:07:30 -07:00
Ishaan Jaff
344010e127 Pass litellm proxy specific metadata 2024-07-23 15:31:30 -07:00
Ishaan Jaff
d6e7cb2df7 fix triton linting 2024-07-23 11:03:34 -07:00
Ishaan Jaff
71c755d9a2 Merge pull request #3905 from giritatavarty-8451/litellm_triton_chatcompletion_support
Litellm triton chatcompletion support - Resubmit of #3895
2024-07-23 10:30:26 -07:00
Krrish Dholakia
dc9dcfefac fix(openai.py): check if error body is a dictionary before indexing in 2024-07-22 18:12:04 -07:00
Krrish Dholakia
548a57b9f6 fix(vertex_httpx.py): Change non-blocking vertex error to warning
Fixes https://github.com/BerriAI/litellm/issues/4825
2024-07-22 14:00:33 -07:00
Krish Dholakia
9ee7ba08b0 Revert "Fix: use Bedrock region from environment variables before other region definitions" 2024-07-22 07:40:24 -07:00
Ishaan Jaff
e43604ef91 fix make_sync_openai_audio_transcriptions_request 2024-07-20 20:03:12 -07:00
Krish Dholakia
b9732b15b7 Merge pull request #4613 from petermuller/main
Fix: use Bedrock region from environment variables before other region definitions
2024-07-20 19:19:05 -07:00
Ishaan Jaff
419c6cd37e fix merge conflicts 2024-07-20 19:14:20 -07:00
Ishaan Jaff
04e3ebf627 Merge branch 'main' into litellm_return-response_headers 2024-07-20 19:05:56 -07:00
Ishaan Jaff
98bc63bff4 fix make_sync_openai_audio_transcriptions_request 2024-07-20 18:17:21 -07:00
Ishaan Jaff
2dcbd5c534 rename to _response_headers 2024-07-20 17:31:16 -07:00
Ishaan Jaff
2e24d17eac return response headers 2024-07-20 15:26:44 -07:00
Krrish Dholakia
a01a45230c fix(openai.py): support completion, streaming, async_streaming 2024-07-20 15:23:42 -07:00
Krrish Dholakia
6a7bf0c251 fix(openai.py): drop invalid params if drop_params: true for azure ai
Fixes https://github.com/BerriAI/litellm/issues/4800
2024-07-20 15:08:15 -07:00
Ishaan Jaff
c07b8d9575 openai - return response headers 2024-07-20 15:04:27 -07:00
Ishaan Jaff
eca6c4ba31 openai return response headers 2024-07-20 14:07:41 -07:00
Krish Dholakia
ae8f785735 Merge pull request #4796 from BerriAI/litellm_refactor_requests_factory
fix(factory.py): refactor factory to use httpx client
2024-07-19 21:07:41 -07:00
Krrish Dholakia
48dd21cc88 fix(anthropic.py): fix streaming client 2024-07-19 18:55:00 -07:00
Krrish Dholakia
399803d3d3 fix(factory.py): refactor factory to use httpx client 2024-07-19 15:35:05 -07:00
Krrish Dholakia
c1e128694f fix(factory.py): refactor factory to use httpx client 2024-07-19 15:32:01 -07:00
Krish Dholakia
3bff1e36f6 Merge branch 'main' into litellm_ai21_jamba 2024-07-19 15:11:55 -07:00
Sophia Loris
adae0777d6 resolve merge conflicts 2024-07-19 09:45:53 -05:00
Sophia Loris
91fa69c0c2 Add support for Triton streaming & triton async completions 2024-07-19 09:35:27 -05:00
Krish Dholakia
b7e6778837 Merge pull request #4773 from titusz/patch-1
Add missing `num_gpu` ollama configuration parameter
2024-07-18 22:47:39 -07:00
Krrish Dholakia
2c9c28e343 fix(anthropic.py): revert client to requests library 2024-07-18 22:45:41 -07:00
Krrish Dholakia
fd9880ebbe feat(vertex_ai_anthropic.py): support response_schema for vertex ai anthropic calls
allows passing response_schema for anthropic calls. supports schema validation.
2024-07-18 22:40:35 -07:00