Krrish Dholakia
|
6a4001c4f4
|
fix(vertex_ai_llama3.py): Fix llama3 streaming issue
Closes https://github.com/BerriAI/litellm/issues/4885
|
2024-07-25 22:30:55 -07:00 |
|
Krish Dholakia
|
a5cea7929d
|
Merge branch 'main' into bedrock-llama3.1-405b
|
2024-07-25 19:29:10 -07:00 |
|
Ishaan Jaff
|
3eea3f0e2d
|
Merge pull request #4884 from Manouchehri/add-mistral-large-2407-bedrock-1
Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock
|
2024-07-25 19:22:46 -07:00 |
|
Krish Dholakia
|
ff6d153d58
|
Merge pull request #4879 from fracapuano/main
Add Single-Token predictions support for Replicate
|
2024-07-25 19:10:57 -07:00 |
|
Krish Dholakia
|
9a42d312b5
|
Merge pull request #4887 from BerriAI/litellm_custom_llm
feat(custom_llm.py): Support Custom LLM Handlers
|
2024-07-25 19:05:29 -07:00 |
|
Krrish Dholakia
|
a2de16582a
|
fix(custom_llm.py): pass input params to custom llm
|
2024-07-25 19:03:52 -07:00 |
|
Ishaan Jaff
|
0b10d40674
|
feat support audio health checks for azure
|
2024-07-25 17:30:15 -07:00 |
|
Ishaan Jaff
|
591857082b
|
feat - support health check audio_speech
|
2024-07-25 17:26:14 -07:00 |
|
Ishaan Jaff
|
56b55b6ab4
|
fix whisper health check with litellm
|
2024-07-25 17:22:57 -07:00 |
|
Krrish Dholakia
|
9b1c7066b7
|
feat(utils.py): support async streaming for custom llm provider
|
2024-07-25 17:11:57 -07:00 |
|
Krrish Dholakia
|
bf23aac11d
|
feat(utils.py): support sync streaming for custom llm provider
|
2024-07-25 16:47:32 -07:00 |
|
Krrish Dholakia
|
fe503386ab
|
fix(custom_llm.py): support async completion calls
|
2024-07-25 15:51:39 -07:00 |
|
Krrish Dholakia
|
54e1ca29b7
|
feat(custom_llm.py): initial working commit for writing your own custom LLM handler
Fixes https://github.com/BerriAI/litellm/issues/4675
Also Addresses https://github.com/BerriAI/litellm/discussions/4677
|
2024-07-25 15:33:05 -07:00 |
|
David Manouchehri
|
213227a7e0
|
Support tool calling for Llama 3.1 on Amazon bedrock.
|
2024-07-25 20:36:25 +00:00 |
|
David Manouchehri
|
c7fd02faf3
|
Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock.
|
2024-07-25 20:04:03 +00:00 |
|
David Manouchehri
|
97d84ce860
|
Add Llama 3.1 405b for Bedrock
|
2024-07-25 19:30:13 +00:00 |
|
fracapuano
|
42ead92211
|
fix: now supports single tokens prediction
|
2024-07-25 19:06:07 +02:00 |
|
Krrish Dholakia
|
b94b9dcdea
|
fix(bedrock_httpx.py): fix async client check
|
2024-07-24 12:43:52 -07:00 |
|
David Manouchehri
|
3ee6baabf7
|
Add Llama 3.1 for Bedrock.
|
2024-07-24 05:48:23 +00:00 |
|
Krish Dholakia
|
29eead1add
|
Merge pull request #4845 from BerriAI/litellm_vertex_ai_llama3_1_api
feat(vertex_ai_llama.py): vertex ai llama3.1 api support
|
2024-07-23 21:51:46 -07:00 |
|
Krrish Dholakia
|
a7779796ef
|
fix(anthropic.py): support openai system message being a list
|
2024-07-23 21:45:56 -07:00 |
|
Krrish Dholakia
|
23a3be184b
|
build(model_prices_and_context_window.json): add model pricing for vertex ai llama 3.1 api
|
2024-07-23 17:36:07 -07:00 |
|
Krrish Dholakia
|
778afcee31
|
feat(vertex_ai_llama.py): vertex ai llama3.1 api support
Initial working commit for vertex ai llama 3.1 api support
|
2024-07-23 17:07:30 -07:00 |
|
Ishaan Jaff
|
344010e127
|
Pass litellm proxy specific metadata
|
2024-07-23 15:31:30 -07:00 |
|
Ishaan Jaff
|
d6e7cb2df7
|
fix triton linting
|
2024-07-23 11:03:34 -07:00 |
|
Ishaan Jaff
|
71c755d9a2
|
Merge pull request #3905 from giritatavarty-8451/litellm_triton_chatcompletion_support
Litellm triton chatcompletion support - Resubmit of #3895
|
2024-07-23 10:30:26 -07:00 |
|
Krrish Dholakia
|
dc9dcfefac
|
fix(openai.py): check if error body is a dictionary before indexing in
|
2024-07-22 18:12:04 -07:00 |
|
Krrish Dholakia
|
548a57b9f6
|
fix(vertex_httpx.py): Change non-blocking vertex error to warning
Fixes https://github.com/BerriAI/litellm/issues/4825
|
2024-07-22 14:00:33 -07:00 |
|
Krish Dholakia
|
9ee7ba08b0
|
Revert "Fix: use Bedrock region from environment variables before other region definitions"
|
2024-07-22 07:40:24 -07:00 |
|
Ishaan Jaff
|
e43604ef91
|
fix make_sync_openai_audio_transcriptions_request
|
2024-07-20 20:03:12 -07:00 |
|
Krish Dholakia
|
b9732b15b7
|
Merge pull request #4613 from petermuller/main
Fix: use Bedrock region from environment variables before other region definitions
|
2024-07-20 19:19:05 -07:00 |
|
Ishaan Jaff
|
419c6cd37e
|
fix merge conflicts
|
2024-07-20 19:14:20 -07:00 |
|
Ishaan Jaff
|
04e3ebf627
|
Merge branch 'main' into litellm_return-response_headers
|
2024-07-20 19:05:56 -07:00 |
|
Ishaan Jaff
|
98bc63bff4
|
fix make_sync_openai_audio_transcriptions_request
|
2024-07-20 18:17:21 -07:00 |
|
Ishaan Jaff
|
2dcbd5c534
|
rename to _response_headers
|
2024-07-20 17:31:16 -07:00 |
|
Ishaan Jaff
|
2e24d17eac
|
return response headers
|
2024-07-20 15:26:44 -07:00 |
|
Krrish Dholakia
|
a01a45230c
|
fix(openai.py): support completion, streaming, async_streaming
|
2024-07-20 15:23:42 -07:00 |
|
Krrish Dholakia
|
6a7bf0c251
|
fix(openai.py): drop invalid params if drop_params: true for azure ai
Fixes https://github.com/BerriAI/litellm/issues/4800
|
2024-07-20 15:08:15 -07:00 |
|
Ishaan Jaff
|
c07b8d9575
|
openai - return response headers
|
2024-07-20 15:04:27 -07:00 |
|
Ishaan Jaff
|
eca6c4ba31
|
openai return response headers
|
2024-07-20 14:07:41 -07:00 |
|
Krish Dholakia
|
ae8f785735
|
Merge pull request #4796 from BerriAI/litellm_refactor_requests_factory
fix(factory.py): refactor factory to use httpx client
|
2024-07-19 21:07:41 -07:00 |
|
Krrish Dholakia
|
48dd21cc88
|
fix(anthropic.py): fix streaming client
|
2024-07-19 18:55:00 -07:00 |
|
Krrish Dholakia
|
399803d3d3
|
fix(factory.py): refactor factory to use httpx client
|
2024-07-19 15:35:05 -07:00 |
|
Krrish Dholakia
|
c1e128694f
|
fix(factory.py): refactor factory to use httpx client
|
2024-07-19 15:32:01 -07:00 |
|
Krish Dholakia
|
3bff1e36f6
|
Merge branch 'main' into litellm_ai21_jamba
|
2024-07-19 15:11:55 -07:00 |
|
Sophia Loris
|
adae0777d6
|
resolve merge conflicts
|
2024-07-19 09:45:53 -05:00 |
|
Sophia Loris
|
91fa69c0c2
|
Add support for Triton streaming & triton async completions
|
2024-07-19 09:35:27 -05:00 |
|
Krish Dholakia
|
b7e6778837
|
Merge pull request #4773 from titusz/patch-1
Add missing `num_gpu` ollama configuration parameter
|
2024-07-18 22:47:39 -07:00 |
|
Krrish Dholakia
|
2c9c28e343
|
fix(anthropic.py): revert client to requests library
|
2024-07-18 22:45:41 -07:00 |
|
Krrish Dholakia
|
fd9880ebbe
|
feat(vertex_ai_anthropic.py): support response_schema for vertex ai anthropic calls
allows passing response_schema for anthropic calls. supports schema validation.
|
2024-07-18 22:40:35 -07:00 |
|