Krrish Dholakia
|
73d8e3e640
|
fix(ollama_chat.py): fix token counting
|
2024-02-06 22:18:46 -08:00 |
|
Krrish Dholakia
|
d1db67890c
|
fix(ollama.py): support format for ollama
|
2024-02-06 10:11:52 -08:00 |
|
Krrish Dholakia
|
9e091a0624
|
fix(ollama_chat.py): explicitly state if ollama call is streaming or not
|
2024-02-06 07:43:47 -08:00 |
|
Krrish Dholakia
|
2e3748e6eb
|
fix(ollama_chat.py): fix ollama chat completion token counting
|
2024-02-06 07:30:26 -08:00 |
|
Ishaan Jaff
|
14c9e239a1
|
Merge pull request #1750 from vanpelt/patch-2
Re-raise exception in async ollama streaming
|
2024-02-05 08:12:17 -08:00 |
|
Krish Dholakia
|
28df60b609
|
Merge pull request #1809 from BerriAI/litellm_embedding_caching_updates
Support caching individual items in embedding list (Async embedding only)
|
2024-02-03 21:04:23 -08:00 |
|
Krrish Dholakia
|
312c7462c8
|
refactor(ollama.py): trigger rebuild
|
2024-02-03 20:23:43 -08:00 |
|
Krrish Dholakia
|
01cef1fe9e
|
fix(ollama.py): fix api connection error
https://github.com/BerriAI/litellm/issues/1735
|
2024-02-03 20:22:33 -08:00 |
|
Krrish Dholakia
|
c49c88c8e5
|
fix(utils.py): route together ai calls to openai client
together ai is now openai-compatible
n
|
2024-02-03 19:22:48 -08:00 |
|
Krish Dholakia
|
6408af11b6
|
Merge pull request #1799 from BerriAI/litellm_bedrock_stable_diffusion_support
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:59:00 -08:00 |
|
Krrish Dholakia
|
36416360c4
|
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:08:38 -08:00 |
|
Krrish Dholakia
|
0ffdf57dec
|
fix(vertex_ai.py): add async embedding support for vertex ai
|
2024-02-03 10:35:17 -08:00 |
|
Krrish Dholakia
|
d9ba8668f4
|
feat(vertex_ai.py): vertex ai gecko text embedding support
|
2024-02-03 09:48:29 -08:00 |
|
Krrish Dholakia
|
0072d796f6
|
fix(vertex_ai.py): fix params
|
2024-02-01 18:09:49 -08:00 |
|
Krrish Dholakia
|
0f9e793daf
|
feat(vertex_ai.py): add support for custom models via vertex ai model garden
|
2024-02-01 17:47:34 -08:00 |
|
Chris Van Pelt
|
1568b162f5
|
Re-raise exception in async ollama streaming
|
2024-02-01 16:14:07 -08:00 |
|
Krrish Dholakia
|
d46df34ff5
|
fix(utils.py): fix streaming chunks to not return role, unless set
|
2024-02-01 09:55:56 -08:00 |
|
Krrish Dholakia
|
37de964da4
|
fix(ollama_chat.py): fix the way optional params are passed in
|
2024-01-30 15:48:48 -08:00 |
|
Ishaan Jaff
|
dd9c78819a
|
Merge pull request #1689 from BerriAI/litellm_set_organization_on_config.yaml
[Feat] Set OpenAI organization for litellm.completion, Proxy Config
|
2024-01-30 11:47:42 -08:00 |
|
ishaan-jaff
|
e011c4a989
|
(fix) use OpenAI organization in ahealth_check
|
2024-01-30 11:45:22 -08:00 |
|
ishaan-jaff
|
ae4e273db7
|
(feat) OpenAI set organization
|
2024-01-30 10:54:56 -08:00 |
|
Ishaan Jaff
|
2686ec0f9f
|
Merge pull request #1685 from BerriAI/litellm_bedrock_emb_input
[Fix] Graceful rejection of token input for AWS Embeddings API
|
2024-01-30 10:32:14 -08:00 |
|
Krish Dholakia
|
c53ad878b3
|
Merge pull request #1657 from eslamkarim/patch-1
Change quota project to the correct project being used for the call
|
2024-01-30 08:40:09 -08:00 |
|
ishaan-jaff
|
f941c57688
|
(fix) use isinstance to check types
|
2024-01-30 08:31:21 -08:00 |
|
ishaan-jaff
|
b9a117fa1d
|
(feat) Bedrock embedding - raise correct exception BadRequest
|
2024-01-30 08:14:35 -08:00 |
|
Krrish Dholakia
|
03961bb8c6
|
fix(factory.py): fix mistral prompt template - doesn't support <<sys>> tokens
|
2024-01-30 07:42:59 -08:00 |
|
Samay Kapadia
|
7d7107cbe1
|
fix mistral's prompt template
|
2024-01-29 12:28:58 +01:00 |
|
Eslam karim gaber
|
2b3965dbf5
|
imports
|
2024-01-28 23:14:06 +02:00 |
|
Eslam karim gaber
|
48dbdaa73e
|
Change quota project to the correct project being used for the call
if not set it will use the default project in the ADC to set that quota project which is usually different
https://github.com/googleapis/python-aiplatform/issues/2557#issuecomment-1709284744
|
2024-01-28 19:55:01 +02:00 |
|
Krish Dholakia
|
ba4089824d
|
Merge pull request #1646 from BerriAI/litellm_image_gen_cost_tracking_proxy
Litellm image gen cost tracking proxy
|
2024-01-26 22:30:14 -08:00 |
|
Krrish Dholakia
|
d755d50901
|
fix(openai.py): fix openai image gen logging
|
2024-01-26 21:05:49 -08:00 |
|
Krrish Dholakia
|
f5da95685a
|
feat(utils.py): support region based pricing for bedrock + use bedrock's token counts if given
|
2024-01-26 14:53:58 -08:00 |
|
Krrish Dholakia
|
09ec6d6458
|
fix(utils.py): fix sagemaker async logging for sync streaming
https://github.com/BerriAI/litellm/issues/1592
|
2024-01-25 12:49:45 -08:00 |
|
Krrish Dholakia
|
43f139fafd
|
fix(ollama_chat.py): fix default token counting for ollama chat
|
2024-01-24 20:09:17 -08:00 |
|
Krrish Dholakia
|
f19f0dad89
|
fix(router.py): fix client init
|
2024-01-22 22:15:39 -08:00 |
|
Ishaan Jaff
|
6d105754d7
|
Merge pull request #1561 from BerriAI/litellm_sagemaker_streaming
[Feat] Add REAL Sagemaker streaming
|
2024-01-22 22:10:20 -08:00 |
|
ishaan-jaff
|
802fd6cebf
|
v0 add TokenIterator, stream support
|
2024-01-22 21:49:44 -08:00 |
|
Krrish Dholakia
|
5e0d99b2ef
|
fix(router.py): fix order of dereferenced dictionaries
|
2024-01-22 21:42:25 -08:00 |
|
Krish Dholakia
|
bedb08bdef
|
Merge pull request #1557 from BerriAI/litellm_emit_spend_logs
feat(utils.py): emit response cost as part of logs
|
2024-01-22 21:02:40 -08:00 |
|
Krrish Dholakia
|
3e8c8ef507
|
fix(openai.py): fix linting issue
|
2024-01-22 18:20:15 -08:00 |
|
Krrish Dholakia
|
06f6f128b6
|
refactor(__init__.py): adds init.py file in tokenizers
|
2024-01-22 15:33:16 -08:00 |
|
Krrish Dholakia
|
29800d1e01
|
fix(azure_dall_e_2.py): handle azure not returning a 'retry-after' param
|
2024-01-22 12:02:02 -08:00 |
|
Krrish Dholakia
|
b07677c6be
|
fix(gemini.py): support streaming
|
2024-01-19 20:21:34 -08:00 |
|
Krrish Dholakia
|
1ec5083542
|
fix(anthropic.py): fix pr for anthropic headers
|
2024-01-19 17:06:24 -08:00 |
|
Keegan McCallum
|
3b719b2afd
|
Allow overriding headers for anthropic
|
2024-01-18 20:12:59 -08:00 |
|
puffo
|
becff369dc
|
fix(ollama_chat.py): use tiktoken as backup for prompt token counting
|
2024-01-18 10:47:24 -06:00 |
|
Krrish Dholakia
|
8e9dc09955
|
fix(bedrock.py): add support for sts based boto3 initialization
https://github.com/BerriAI/litellm/issues/1476
|
2024-01-17 12:08:59 -08:00 |
|
ishaan-jaff
|
5a8a5fa0fd
|
(fix) using base_url Azure
|
2024-01-17 10:12:55 -08:00 |
|
ishaan-jaff
|
7178d01c8f
|
(feat) support base_url with /openai passed for Azure
|
2024-01-17 10:03:25 -08:00 |
|
ishaan-jaff
|
b95d6ec207
|
(v0) fixes for Azure GPT Vision enhancements
|
2024-01-17 09:57:16 -08:00 |
|