Commit graph

536 commits

Author SHA1 Message Date
Krish Dholakia
57654f4533
Merge branch 'main' into litellm_aioboto3_sagemaker 2024-02-14 21:46:58 -08:00
Krrish Dholakia
fe1fe70c64 fix(vertex_ai.py): map finish reason 2024-02-14 11:42:13 -08:00
Krrish Dholakia
cb5a13ed49 fix(bedrock.py): fix amazon titan prompt formatting 2024-02-13 22:02:25 -08:00
Krrish Dholakia
3ef391800a fix(sagemaker.py): fix token iterator default flag 2024-02-13 21:41:09 -08:00
Krrish Dholakia
f09c09ace4 docs(pii_masking.md): fix presidio tutorial 2024-02-13 07:42:27 -08:00
Krrish Dholakia
2f815705ca fix(sagemaker.py): use __anext__ 2024-02-12 22:13:35 -08:00
Krrish Dholakia
b1bc30ee16 feat(sagemaker.py): aioboto3 streaming support 2024-02-12 21:18:34 -08:00
Krrish Dholakia
460b48914e feat(sagemaker.py): initial commit of working sagemaker with aioboto3 2024-02-12 17:25:57 -08:00
ishaan-jaff
896fd393db (feat) support bedrock timeout 2024-02-09 14:36:43 -08:00
Krish Dholakia
51c07e294a
Merge pull request #1902 from BerriAI/litellm_mistral_message_list_fix
fix(factory.py): mistral message input fix
2024-02-08 23:01:39 -08:00
Krrish Dholakia
841639333b fix(bedrock.py): raise exception for amazon titan null response 2024-02-08 21:12:25 -08:00
Krrish Dholakia
c9e5c796ad fix(factory.py): mistral message input fix 2024-02-08 20:54:26 -08:00
David Leen
140d915adf Add support for AWS credentials from profile file
https://boto3.amazonaws.com/v1/documentation/api/latest/guide/credentials.html#aws-config-file
2024-02-08 15:10:50 -08:00
Krrish Dholakia
73d8e3e640 fix(ollama_chat.py): fix token counting 2024-02-06 22:18:46 -08:00
Krrish Dholakia
d1db67890c fix(ollama.py): support format for ollama 2024-02-06 10:11:52 -08:00
Krrish Dholakia
9e091a0624 fix(ollama_chat.py): explicitly state if ollama call is streaming or not 2024-02-06 07:43:47 -08:00
Krrish Dholakia
2e3748e6eb fix(ollama_chat.py): fix ollama chat completion token counting 2024-02-06 07:30:26 -08:00
Ishaan Jaff
14c9e239a1
Merge pull request #1750 from vanpelt/patch-2
Re-raise exception in async ollama streaming
2024-02-05 08:12:17 -08:00
Krish Dholakia
28df60b609
Merge pull request #1809 from BerriAI/litellm_embedding_caching_updates
Support caching individual items in embedding list (Async embedding only)
2024-02-03 21:04:23 -08:00
Krrish Dholakia
312c7462c8 refactor(ollama.py): trigger rebuild 2024-02-03 20:23:43 -08:00
Krrish Dholakia
01cef1fe9e fix(ollama.py): fix api connection error
https://github.com/BerriAI/litellm/issues/1735
2024-02-03 20:22:33 -08:00
Krrish Dholakia
c49c88c8e5 fix(utils.py): route together ai calls to openai client
together ai is now openai-compatible

n
2024-02-03 19:22:48 -08:00
Krish Dholakia
6408af11b6
Merge pull request #1799 from BerriAI/litellm_bedrock_stable_diffusion_support
feat(bedrock.py): add stable diffusion image generation support
2024-02-03 12:59:00 -08:00
Krrish Dholakia
36416360c4 feat(bedrock.py): add stable diffusion image generation support 2024-02-03 12:08:38 -08:00
Krrish Dholakia
0ffdf57dec fix(vertex_ai.py): add async embedding support for vertex ai 2024-02-03 10:35:17 -08:00
Krrish Dholakia
d9ba8668f4 feat(vertex_ai.py): vertex ai gecko text embedding support 2024-02-03 09:48:29 -08:00
Krrish Dholakia
0072d796f6 fix(vertex_ai.py): fix params 2024-02-01 18:09:49 -08:00
Krrish Dholakia
0f9e793daf feat(vertex_ai.py): add support for custom models via vertex ai model garden 2024-02-01 17:47:34 -08:00
Chris Van Pelt
1568b162f5
Re-raise exception in async ollama streaming 2024-02-01 16:14:07 -08:00
Krrish Dholakia
d46df34ff5 fix(utils.py): fix streaming chunks to not return role, unless set 2024-02-01 09:55:56 -08:00
Krrish Dholakia
37de964da4 fix(ollama_chat.py): fix the way optional params are passed in 2024-01-30 15:48:48 -08:00
Ishaan Jaff
dd9c78819a
Merge pull request #1689 from BerriAI/litellm_set_organization_on_config.yaml
[Feat] Set OpenAI organization for litellm.completion, Proxy Config
2024-01-30 11:47:42 -08:00
ishaan-jaff
e011c4a989 (fix) use OpenAI organization in ahealth_check 2024-01-30 11:45:22 -08:00
ishaan-jaff
ae4e273db7 (feat) OpenAI set organization 2024-01-30 10:54:56 -08:00
Ishaan Jaff
2686ec0f9f
Merge pull request #1685 from BerriAI/litellm_bedrock_emb_input
[Fix] Graceful rejection of token input for AWS Embeddings API
2024-01-30 10:32:14 -08:00
Krish Dholakia
c53ad878b3
Merge pull request #1657 from eslamkarim/patch-1
Change quota project to the correct project being used for the call
2024-01-30 08:40:09 -08:00
ishaan-jaff
f941c57688 (fix) use isinstance to check types 2024-01-30 08:31:21 -08:00
ishaan-jaff
b9a117fa1d (feat) Bedrock embedding - raise correct exception BadRequest 2024-01-30 08:14:35 -08:00
Krrish Dholakia
03961bb8c6 fix(factory.py): fix mistral prompt template - doesn't support <<sys>> tokens 2024-01-30 07:42:59 -08:00
Samay Kapadia
7d7107cbe1
fix mistral's prompt template 2024-01-29 12:28:58 +01:00
Eslam karim gaber
2b3965dbf5
imports 2024-01-28 23:14:06 +02:00
Eslam karim gaber
48dbdaa73e
Change quota project to the correct project being used for the call
if not set it will use the default project in the ADC to set that quota project which is usually different 
https://github.com/googleapis/python-aiplatform/issues/2557#issuecomment-1709284744
2024-01-28 19:55:01 +02:00
Krish Dholakia
ba4089824d
Merge pull request #1646 from BerriAI/litellm_image_gen_cost_tracking_proxy
Litellm image gen cost tracking proxy
2024-01-26 22:30:14 -08:00
Krrish Dholakia
d755d50901 fix(openai.py): fix openai image gen logging 2024-01-26 21:05:49 -08:00
Krrish Dholakia
f5da95685a feat(utils.py): support region based pricing for bedrock + use bedrock's token counts if given 2024-01-26 14:53:58 -08:00
Krrish Dholakia
09ec6d6458 fix(utils.py): fix sagemaker async logging for sync streaming
https://github.com/BerriAI/litellm/issues/1592
2024-01-25 12:49:45 -08:00
Krrish Dholakia
43f139fafd fix(ollama_chat.py): fix default token counting for ollama chat 2024-01-24 20:09:17 -08:00
Krrish Dholakia
f19f0dad89 fix(router.py): fix client init 2024-01-22 22:15:39 -08:00
Ishaan Jaff
6d105754d7
Merge pull request #1561 from BerriAI/litellm_sagemaker_streaming
[Feat] Add REAL Sagemaker streaming
2024-01-22 22:10:20 -08:00
ishaan-jaff
802fd6cebf v0 add TokenIterator, stream support 2024-01-22 21:49:44 -08:00