Commit graph

935 commits

Author SHA1 Message Date
Mateo Cámara
bb06c51ede Added test to check if acompletion is using the same parameters as CompletionRequest attributes. Added functools to client decorator to expose acompletion parameters from outside. 2024-01-09 12:06:49 +01:00
ishaan-jaff
09874cc83f (v0) add ContentPolicyViolationError 2024-01-09 16:33:03 +05:30
ishaan-jaff
5f2cbfc711 (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
dd78782133 fix(utils.py): error handling for litellm --model mistral edge case 2024-01-08 15:09:01 +05:30
Krrish Dholakia
6333fbfe56 fix(main.py): support cost calculation for text completion streaming object 2024-01-08 12:41:43 +05:30
Krrish Dholakia
9b46412279 fix(utils.py): fix logging for text completion streaming 2024-01-08 12:05:28 +05:30
Krrish Dholakia
c04fa54d19 fix(utils.py): fix exception raised 2024-01-08 07:42:17 +05:30
Krrish Dholakia
3469b5b911 fix(utils.py): map optional params for gemini 2024-01-08 07:38:55 +05:30
Krrish Dholakia
75177c2a15 bump: version 1.16.16 → 1.16.17 2024-01-08 07:16:37 +05:30
Krish Dholakia
439ee3bafc
Merge pull request #1344 from BerriAI/litellm_speed_improvements
Litellm speed improvements
2024-01-06 22:38:10 +05:30
Krrish Dholakia
5fd2f945f3 fix(factory.py): support gemini-pro-vision on google ai studio
https://github.com/BerriAI/litellm/issues/1329
2024-01-06 22:36:22 +05:30
Krrish Dholakia
712f89b4f1 fix(utils.py): handle original_response being a json 2024-01-06 17:02:50 +05:30
ishaan-jaff
4679c7b99a (fix) caching use same "created" in response_object 2024-01-05 16:03:56 +05:30
ishaan-jaff
00b001b96b (feat) completion_cost: improve model=None error 2024-01-05 15:26:04 +05:30
ishaan-jaff
f681f0f2b2 (feat) completion_cost - embeddings + raise Exception 2024-01-05 13:11:23 +05:30
Krrish Dholakia
aa72d65c90 fix(utils.py): fix check for if cached response should be returned 2024-01-04 21:49:19 +05:30
Krrish Dholakia
773a0a147a fix(utils.py): raise a bad request error if litellm client raises a model /provider not found error 2024-01-04 15:50:43 +05:30
Krrish Dholakia
99d9a825de feat(proxy_server.py): abstract config update/writing and support persisting config in db
allows user to opt into writing to db (SAVE_CONFIG_TO_DB) and removes any api keys before sending to db

 https://github.com/BerriAI/litellm/issues/1322
2024-01-04 14:44:59 +05:30
Krrish Dholakia
8cee267a5b fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
2024-01-03 12:42:43 +05:30
ishaan-jaff
96cb6f3b10 (fix) azure+stream: count completion tokens 2024-01-03 12:06:39 +05:30
ishaan-jaff
f3b8d9c3ef (fix) counting response tokens+streaming 2024-01-03 12:06:39 +05:30
Krrish Dholakia
a778f8a00e bump: version 1.16.10 → 1.16.11 2024-01-02 22:26:47 +05:30
ishaan-jaff
790dcff5e0 (feat) add xinference as an embedding provider 2024-01-02 15:32:26 +05:30
Krrish Dholakia
0fffcc1579 fix(utils.py): support token counting for gpt-4-vision models 2024-01-02 14:41:42 +05:30
Krrish Dholakia
a83e2e07cf fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
2024-01-01 21:22:42 +05:30
ishaan-jaff
535a547b66 (fix) use cloudflare optional params 2023-12-30 12:22:31 +05:30
Krrish Dholakia
77be3e3114 fix(main.py): don't set timeout as an optional api param 2023-12-30 11:47:07 +05:30
Krrish Dholakia
38f55249e1 fix(router.py): support retry and fallbacks for atext_completion 2023-12-30 11:19:32 +05:30
ishaan-jaff
5d6954895f (fix) timeout optional param 2023-12-30 11:07:52 +05:30
ishaan-jaff
2f4cd3b569 (feat) proxy - support dynamic timeout per request 2023-12-30 10:55:42 +05:30
Marmik Pandya
1426594d3f add support for mistral json mode via anyscale 2023-12-29 22:26:22 +05:30
ishaan-jaff
e6a7212d10 (fix) counting streaming prompt tokens - azure 2023-12-29 16:13:52 +05:30
ishaan-jaff
8c03be59a8 (fix) token_counter for tool calling 2023-12-29 15:54:03 +05:30
ishaan-jaff
037dcbbe10 (fix) use openai token counter for azure llms 2023-12-29 15:37:46 +05:30
ishaan-jaff
a300ab9152 (feat) azure stream - count correct prompt tokens 2023-12-29 15:15:39 +05:30
ishaan-jaff
3973b9c8e4 (feat) cloudflare - add exception mapping 2023-12-29 12:31:10 +05:30
ishaan-jaff
ee682be093 (feat) add cloudflare streaming 2023-12-29 12:01:26 +05:30
ishaan-jaff
dde6bc4fb6 (feat) cloudflare - add optional params 2023-12-29 11:50:09 +05:30
ishaan-jaff
daf32f3bd4 (fix) tg AI cost tracking - zero-one-ai/Yi-34B-Chat 2023-12-29 09:14:07 +05:30
ishaan-jaff
d79df3a1e9 (fix) together_ai cost tracking 2023-12-28 22:11:08 +05:30
Krrish Dholakia
86403cd14e fix(vertex_ai.py): support function calling for gemini 2023-12-28 19:07:04 +05:30
ishaan-jaff
95e6d2fbba (feat) add voyage ai embeddings 2023-12-28 17:10:15 +05:30
Krrish Dholakia
c4fc28ab0d fix(utils.py): use local tiktoken copy 2023-12-28 11:22:33 +05:30
Krrish Dholakia
e516cfe9f5 fix(utils.py): allow text completion input to be either model or engine 2023-12-27 17:24:16 +05:30
ishaan-jaff
7864d9a027 (fix) openai + stream - logprobs check 2023-12-27 16:59:56 +05:30
ishaan-jaff
646c106983 (feat) text-completion-openai, send 1 finish_reason 2023-12-27 15:45:40 +05:30
ishaan-jaff
e70f588b87 (fix) text_completion use correct finish reason 2023-12-27 15:20:26 +05:30
Krrish Dholakia
c9fdbaf898 fix(azure.py,-openai.py): correctly raise errors if streaming calls fail 2023-12-27 15:08:37 +05:30
Krrish Dholakia
9ba520cc8b fix(google_kms.py): support enums for key management system 2023-12-27 13:19:33 +05:30
ishaan-jaff
4cc59d21d0 (feat) add text_completion, atext_completion CallTypes 2023-12-27 12:24:16 +05:30