Krrish Dholakia
|
dd78782133
|
fix(utils.py): error handling for litellm --model mistral edge case
|
2024-01-08 15:09:01 +05:30 |
|
Krrish Dholakia
|
6333fbfe56
|
fix(main.py): support cost calculation for text completion streaming object
|
2024-01-08 12:41:43 +05:30 |
|
Krrish Dholakia
|
9b46412279
|
fix(utils.py): fix logging for text completion streaming
|
2024-01-08 12:05:28 +05:30 |
|
Krrish Dholakia
|
c04fa54d19
|
fix(utils.py): fix exception raised
|
2024-01-08 07:42:17 +05:30 |
|
Krrish Dholakia
|
3469b5b911
|
fix(utils.py): map optional params for gemini
|
2024-01-08 07:38:55 +05:30 |
|
Krrish Dholakia
|
75177c2a15
|
bump: version 1.16.16 → 1.16.17
|
2024-01-08 07:16:37 +05:30 |
|
Krish Dholakia
|
439ee3bafc
|
Merge pull request #1344 from BerriAI/litellm_speed_improvements
Litellm speed improvements
|
2024-01-06 22:38:10 +05:30 |
|
Krrish Dholakia
|
5fd2f945f3
|
fix(factory.py): support gemini-pro-vision on google ai studio
https://github.com/BerriAI/litellm/issues/1329
|
2024-01-06 22:36:22 +05:30 |
|
Krrish Dholakia
|
712f89b4f1
|
fix(utils.py): handle original_response being a json
|
2024-01-06 17:02:50 +05:30 |
|
ishaan-jaff
|
4679c7b99a
|
(fix) caching use same "created" in response_object
|
2024-01-05 16:03:56 +05:30 |
|
ishaan-jaff
|
00b001b96b
|
(feat) completion_cost: improve model=None error
|
2024-01-05 15:26:04 +05:30 |
|
ishaan-jaff
|
f681f0f2b2
|
(feat) completion_cost - embeddings + raise Exception
|
2024-01-05 13:11:23 +05:30 |
|
Krrish Dholakia
|
aa72d65c90
|
fix(utils.py): fix check for if cached response should be returned
|
2024-01-04 21:49:19 +05:30 |
|
Krrish Dholakia
|
773a0a147a
|
fix(utils.py): raise a bad request error if litellm client raises a model /provider not found error
|
2024-01-04 15:50:43 +05:30 |
|
Krrish Dholakia
|
99d9a825de
|
feat(proxy_server.py): abstract config update/writing and support persisting config in db
allows user to opt into writing to db (SAVE_CONFIG_TO_DB) and removes any api keys before sending to db
https://github.com/BerriAI/litellm/issues/1322
|
2024-01-04 14:44:59 +05:30 |
|
Krrish Dholakia
|
8cee267a5b
|
fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
|
2024-01-03 12:42:43 +05:30 |
|
ishaan-jaff
|
96cb6f3b10
|
(fix) azure+stream: count completion tokens
|
2024-01-03 12:06:39 +05:30 |
|
ishaan-jaff
|
f3b8d9c3ef
|
(fix) counting response tokens+streaming
|
2024-01-03 12:06:39 +05:30 |
|
Krrish Dholakia
|
a778f8a00e
|
bump: version 1.16.10 → 1.16.11
|
2024-01-02 22:26:47 +05:30 |
|
ishaan-jaff
|
790dcff5e0
|
(feat) add xinference as an embedding provider
|
2024-01-02 15:32:26 +05:30 |
|
Krrish Dholakia
|
0fffcc1579
|
fix(utils.py): support token counting for gpt-4-vision models
|
2024-01-02 14:41:42 +05:30 |
|
Krrish Dholakia
|
a83e2e07cf
|
fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
|
2024-01-01 21:22:42 +05:30 |
|
ishaan-jaff
|
535a547b66
|
(fix) use cloudflare optional params
|
2023-12-30 12:22:31 +05:30 |
|
Krrish Dholakia
|
77be3e3114
|
fix(main.py): don't set timeout as an optional api param
|
2023-12-30 11:47:07 +05:30 |
|
Krrish Dholakia
|
38f55249e1
|
fix(router.py): support retry and fallbacks for atext_completion
|
2023-12-30 11:19:32 +05:30 |
|
ishaan-jaff
|
5d6954895f
|
(fix) timeout optional param
|
2023-12-30 11:07:52 +05:30 |
|
ishaan-jaff
|
2f4cd3b569
|
(feat) proxy - support dynamic timeout per request
|
2023-12-30 10:55:42 +05:30 |
|
Marmik Pandya
|
1426594d3f
|
add support for mistral json mode via anyscale
|
2023-12-29 22:26:22 +05:30 |
|
ishaan-jaff
|
e6a7212d10
|
(fix) counting streaming prompt tokens - azure
|
2023-12-29 16:13:52 +05:30 |
|
ishaan-jaff
|
8c03be59a8
|
(fix) token_counter for tool calling
|
2023-12-29 15:54:03 +05:30 |
|
ishaan-jaff
|
037dcbbe10
|
(fix) use openai token counter for azure llms
|
2023-12-29 15:37:46 +05:30 |
|
ishaan-jaff
|
a300ab9152
|
(feat) azure stream - count correct prompt tokens
|
2023-12-29 15:15:39 +05:30 |
|
ishaan-jaff
|
3973b9c8e4
|
(feat) cloudflare - add exception mapping
|
2023-12-29 12:31:10 +05:30 |
|
ishaan-jaff
|
ee682be093
|
(feat) add cloudflare streaming
|
2023-12-29 12:01:26 +05:30 |
|
ishaan-jaff
|
dde6bc4fb6
|
(feat) cloudflare - add optional params
|
2023-12-29 11:50:09 +05:30 |
|
ishaan-jaff
|
daf32f3bd4
|
(fix) tg AI cost tracking - zero-one-ai/Yi-34B-Chat
|
2023-12-29 09:14:07 +05:30 |
|
ishaan-jaff
|
d79df3a1e9
|
(fix) together_ai cost tracking
|
2023-12-28 22:11:08 +05:30 |
|
Krrish Dholakia
|
86403cd14e
|
fix(vertex_ai.py): support function calling for gemini
|
2023-12-28 19:07:04 +05:30 |
|
ishaan-jaff
|
95e6d2fbba
|
(feat) add voyage ai embeddings
|
2023-12-28 17:10:15 +05:30 |
|
Krrish Dholakia
|
c4fc28ab0d
|
fix(utils.py): use local tiktoken copy
|
2023-12-28 11:22:33 +05:30 |
|
Krrish Dholakia
|
e516cfe9f5
|
fix(utils.py): allow text completion input to be either model or engine
|
2023-12-27 17:24:16 +05:30 |
|
ishaan-jaff
|
7864d9a027
|
(fix) openai + stream - logprobs check
|
2023-12-27 16:59:56 +05:30 |
|
ishaan-jaff
|
646c106983
|
(feat) text-completion-openai, send 1 finish_reason
|
2023-12-27 15:45:40 +05:30 |
|
ishaan-jaff
|
e70f588b87
|
(fix) text_completion use correct finish reason
|
2023-12-27 15:20:26 +05:30 |
|
Krrish Dholakia
|
c9fdbaf898
|
fix(azure.py,-openai.py): correctly raise errors if streaming calls fail
|
2023-12-27 15:08:37 +05:30 |
|
Krrish Dholakia
|
9ba520cc8b
|
fix(google_kms.py): support enums for key management system
|
2023-12-27 13:19:33 +05:30 |
|
ishaan-jaff
|
4cc59d21d0
|
(feat) add text_completion, atext_completion CallTypes
|
2023-12-27 12:24:16 +05:30 |
|
ishaan-jaff
|
ef074c707a
|
(fix) streaming logprobs=None
|
2023-12-26 15:42:51 +05:30 |
|
Krrish Dholakia
|
2070a785a4
|
feat(utils.py): support google kms for secret management
https://github.com/BerriAI/litellm/issues/1235
|
2023-12-26 15:39:40 +05:30 |
|
ishaan-jaff
|
6406046d38
|
(feat) logprobs for streaming openai
|
2023-12-26 15:15:05 +05:30 |
|