Krish Dholakia
|
57654f4533
|
Merge branch 'main' into litellm_aioboto3_sagemaker
|
2024-02-14 21:46:58 -08:00 |
|
Krrish Dholakia
|
fe1fe70c64
|
fix(vertex_ai.py): map finish reason
|
2024-02-14 11:42:13 -08:00 |
|
Krrish Dholakia
|
cb5a13ed49
|
fix(bedrock.py): fix amazon titan prompt formatting
|
2024-02-13 22:02:25 -08:00 |
|
Krrish Dholakia
|
3ef391800a
|
fix(sagemaker.py): fix token iterator default flag
|
2024-02-13 21:41:09 -08:00 |
|
Krrish Dholakia
|
f09c09ace4
|
docs(pii_masking.md): fix presidio tutorial
|
2024-02-13 07:42:27 -08:00 |
|
Krrish Dholakia
|
2f815705ca
|
fix(sagemaker.py): use __anext__
|
2024-02-12 22:13:35 -08:00 |
|
Krrish Dholakia
|
b1bc30ee16
|
feat(sagemaker.py): aioboto3 streaming support
|
2024-02-12 21:18:34 -08:00 |
|
Krrish Dholakia
|
460b48914e
|
feat(sagemaker.py): initial commit of working sagemaker with aioboto3
|
2024-02-12 17:25:57 -08:00 |
|
ishaan-jaff
|
896fd393db
|
(feat) support bedrock timeout
|
2024-02-09 14:36:43 -08:00 |
|
Krish Dholakia
|
51c07e294a
|
Merge pull request #1902 from BerriAI/litellm_mistral_message_list_fix
fix(factory.py): mistral message input fix
|
2024-02-08 23:01:39 -08:00 |
|
Krrish Dholakia
|
841639333b
|
fix(bedrock.py): raise exception for amazon titan null response
|
2024-02-08 21:12:25 -08:00 |
|
Krrish Dholakia
|
c9e5c796ad
|
fix(factory.py): mistral message input fix
|
2024-02-08 20:54:26 -08:00 |
|
David Leen
|
140d915adf
|
Add support for AWS credentials from profile file
https://boto3.amazonaws.com/v1/documentation/api/latest/guide/credentials.html#aws-config-file
|
2024-02-08 15:10:50 -08:00 |
|
Krrish Dholakia
|
73d8e3e640
|
fix(ollama_chat.py): fix token counting
|
2024-02-06 22:18:46 -08:00 |
|
Krrish Dholakia
|
d1db67890c
|
fix(ollama.py): support format for ollama
|
2024-02-06 10:11:52 -08:00 |
|
Krrish Dholakia
|
9e091a0624
|
fix(ollama_chat.py): explicitly state if ollama call is streaming or not
|
2024-02-06 07:43:47 -08:00 |
|
Krrish Dholakia
|
2e3748e6eb
|
fix(ollama_chat.py): fix ollama chat completion token counting
|
2024-02-06 07:30:26 -08:00 |
|
Ishaan Jaff
|
14c9e239a1
|
Merge pull request #1750 from vanpelt/patch-2
Re-raise exception in async ollama streaming
|
2024-02-05 08:12:17 -08:00 |
|
Krish Dholakia
|
28df60b609
|
Merge pull request #1809 from BerriAI/litellm_embedding_caching_updates
Support caching individual items in embedding list (Async embedding only)
|
2024-02-03 21:04:23 -08:00 |
|
Krrish Dholakia
|
312c7462c8
|
refactor(ollama.py): trigger rebuild
|
2024-02-03 20:23:43 -08:00 |
|
Krrish Dholakia
|
01cef1fe9e
|
fix(ollama.py): fix api connection error
https://github.com/BerriAI/litellm/issues/1735
|
2024-02-03 20:22:33 -08:00 |
|
Krrish Dholakia
|
c49c88c8e5
|
fix(utils.py): route together ai calls to openai client
together ai is now openai-compatible
n
|
2024-02-03 19:22:48 -08:00 |
|
Krish Dholakia
|
6408af11b6
|
Merge pull request #1799 from BerriAI/litellm_bedrock_stable_diffusion_support
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:59:00 -08:00 |
|
Krrish Dholakia
|
36416360c4
|
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:08:38 -08:00 |
|
Krrish Dholakia
|
0ffdf57dec
|
fix(vertex_ai.py): add async embedding support for vertex ai
|
2024-02-03 10:35:17 -08:00 |
|
Krrish Dholakia
|
d9ba8668f4
|
feat(vertex_ai.py): vertex ai gecko text embedding support
|
2024-02-03 09:48:29 -08:00 |
|
Krrish Dholakia
|
0072d796f6
|
fix(vertex_ai.py): fix params
|
2024-02-01 18:09:49 -08:00 |
|
Krrish Dholakia
|
0f9e793daf
|
feat(vertex_ai.py): add support for custom models via vertex ai model garden
|
2024-02-01 17:47:34 -08:00 |
|
Chris Van Pelt
|
1568b162f5
|
Re-raise exception in async ollama streaming
|
2024-02-01 16:14:07 -08:00 |
|
Krrish Dholakia
|
d46df34ff5
|
fix(utils.py): fix streaming chunks to not return role, unless set
|
2024-02-01 09:55:56 -08:00 |
|
Krrish Dholakia
|
37de964da4
|
fix(ollama_chat.py): fix the way optional params are passed in
|
2024-01-30 15:48:48 -08:00 |
|
Ishaan Jaff
|
dd9c78819a
|
Merge pull request #1689 from BerriAI/litellm_set_organization_on_config.yaml
[Feat] Set OpenAI organization for litellm.completion, Proxy Config
|
2024-01-30 11:47:42 -08:00 |
|
ishaan-jaff
|
e011c4a989
|
(fix) use OpenAI organization in ahealth_check
|
2024-01-30 11:45:22 -08:00 |
|
ishaan-jaff
|
ae4e273db7
|
(feat) OpenAI set organization
|
2024-01-30 10:54:56 -08:00 |
|
Ishaan Jaff
|
2686ec0f9f
|
Merge pull request #1685 from BerriAI/litellm_bedrock_emb_input
[Fix] Graceful rejection of token input for AWS Embeddings API
|
2024-01-30 10:32:14 -08:00 |
|
Krish Dholakia
|
c53ad878b3
|
Merge pull request #1657 from eslamkarim/patch-1
Change quota project to the correct project being used for the call
|
2024-01-30 08:40:09 -08:00 |
|
ishaan-jaff
|
f941c57688
|
(fix) use isinstance to check types
|
2024-01-30 08:31:21 -08:00 |
|
ishaan-jaff
|
b9a117fa1d
|
(feat) Bedrock embedding - raise correct exception BadRequest
|
2024-01-30 08:14:35 -08:00 |
|
Krrish Dholakia
|
03961bb8c6
|
fix(factory.py): fix mistral prompt template - doesn't support <<sys>> tokens
|
2024-01-30 07:42:59 -08:00 |
|
Samay Kapadia
|
7d7107cbe1
|
fix mistral's prompt template
|
2024-01-29 12:28:58 +01:00 |
|
Eslam karim gaber
|
2b3965dbf5
|
imports
|
2024-01-28 23:14:06 +02:00 |
|
Eslam karim gaber
|
48dbdaa73e
|
Change quota project to the correct project being used for the call
if not set it will use the default project in the ADC to set that quota project which is usually different
https://github.com/googleapis/python-aiplatform/issues/2557#issuecomment-1709284744
|
2024-01-28 19:55:01 +02:00 |
|
Krish Dholakia
|
ba4089824d
|
Merge pull request #1646 from BerriAI/litellm_image_gen_cost_tracking_proxy
Litellm image gen cost tracking proxy
|
2024-01-26 22:30:14 -08:00 |
|
Krrish Dholakia
|
d755d50901
|
fix(openai.py): fix openai image gen logging
|
2024-01-26 21:05:49 -08:00 |
|
Krrish Dholakia
|
f5da95685a
|
feat(utils.py): support region based pricing for bedrock + use bedrock's token counts if given
|
2024-01-26 14:53:58 -08:00 |
|
Krrish Dholakia
|
09ec6d6458
|
fix(utils.py): fix sagemaker async logging for sync streaming
https://github.com/BerriAI/litellm/issues/1592
|
2024-01-25 12:49:45 -08:00 |
|
Krrish Dholakia
|
43f139fafd
|
fix(ollama_chat.py): fix default token counting for ollama chat
|
2024-01-24 20:09:17 -08:00 |
|
Krrish Dholakia
|
f19f0dad89
|
fix(router.py): fix client init
|
2024-01-22 22:15:39 -08:00 |
|
Ishaan Jaff
|
6d105754d7
|
Merge pull request #1561 from BerriAI/litellm_sagemaker_streaming
[Feat] Add REAL Sagemaker streaming
|
2024-01-22 22:10:20 -08:00 |
|
ishaan-jaff
|
802fd6cebf
|
v0 add TokenIterator, stream support
|
2024-01-22 21:49:44 -08:00 |
|