Commit graph

935 commits

Author SHA1 Message Date
Mateo Cámara
fb4b5eb4c1 Added test to check if acompletion is using the same parameters as CompletionRequest attributes. Added functools to client decorator to expose acompletion parameters from outside. 2024-01-09 12:06:49 +01:00
ishaan-jaff
66b23ecbb5 (v0) add ContentPolicyViolationError 2024-01-09 16:33:03 +05:30
ishaan-jaff
3081dc525a (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
5a7d2d9c80 fix(utils.py): error handling for litellm --model mistral edge case 2024-01-08 15:09:01 +05:30
Krrish Dholakia
5daa3ce237 fix(main.py): support cost calculation for text completion streaming object 2024-01-08 12:41:43 +05:30
Krrish Dholakia
f79113a7ac fix(utils.py): fix logging for text completion streaming 2024-01-08 12:05:28 +05:30
Krrish Dholakia
9cbb796874 fix(utils.py): fix exception raised 2024-01-08 07:42:17 +05:30
Krrish Dholakia
d8a2675643 fix(utils.py): map optional params for gemini 2024-01-08 07:38:55 +05:30
Krrish Dholakia
4c385a8cf5 bump: version 1.16.16 → 1.16.17 2024-01-08 07:16:37 +05:30
Krish Dholakia
67ecab4b38 Merge pull request #1344 from BerriAI/litellm_speed_improvements
Litellm speed improvements
2024-01-06 22:38:10 +05:30
Krrish Dholakia
2d1871a1ae fix(factory.py): support gemini-pro-vision on google ai studio
https://github.com/BerriAI/litellm/issues/1329
2024-01-06 22:36:22 +05:30
Krrish Dholakia
9c4cc06e64 fix(utils.py): handle original_response being a json 2024-01-06 17:02:50 +05:30
ishaan-jaff
eb8dba100d (fix) caching use same "created" in response_object 2024-01-05 16:03:56 +05:30
ishaan-jaff
9dc0074613 (feat) completion_cost: improve model=None error 2024-01-05 15:26:04 +05:30
ishaan-jaff
9313bda4c8 (feat) completion_cost - embeddings + raise Exception 2024-01-05 13:11:23 +05:30
Krrish Dholakia
52b78dc074 fix(utils.py): fix check for if cached response should be returned 2024-01-04 21:49:19 +05:30
Krrish Dholakia
6f7c764d5e fix(utils.py): raise a bad request error if litellm client raises a model /provider not found error 2024-01-04 15:50:43 +05:30
Krrish Dholakia
5aecbced99 feat(proxy_server.py): abstract config update/writing and support persisting config in db
allows user to opt into writing to db (SAVE_CONFIG_TO_DB) and removes any api keys before sending to db

 https://github.com/BerriAI/litellm/issues/1322
2024-01-04 14:44:59 +05:30
Krrish Dholakia
176af67aac fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
2024-01-03 12:42:43 +05:30
ishaan-jaff
f67b56ddf5 (fix) azure+stream: count completion tokens 2024-01-03 12:06:39 +05:30
ishaan-jaff
f582ef666f (fix) counting response tokens+streaming 2024-01-03 12:06:39 +05:30
Krrish Dholakia
adef0d9410 bump: version 1.16.10 → 1.16.11 2024-01-02 22:26:47 +05:30
ishaan-jaff
0e8809abf2 (feat) add xinference as an embedding provider 2024-01-02 15:32:26 +05:30
Krrish Dholakia
d4da63800e fix(utils.py): support token counting for gpt-4-vision models 2024-01-02 14:41:42 +05:30
Krrish Dholakia
4eae0c9a0d fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
2024-01-01 21:22:42 +05:30
ishaan-jaff
31bdcb48af (fix) use cloudflare optional params 2023-12-30 12:22:31 +05:30
Krrish Dholakia
7d55a563ee fix(main.py): don't set timeout as an optional api param 2023-12-30 11:47:07 +05:30
Krrish Dholakia
e1925d0e29 fix(router.py): support retry and fallbacks for atext_completion 2023-12-30 11:19:32 +05:30
ishaan-jaff
fa4a533e91 (fix) timeout optional param 2023-12-30 11:07:52 +05:30
ishaan-jaff
d5cbef4e36 (feat) proxy - support dynamic timeout per request 2023-12-30 10:55:42 +05:30
Marmik Pandya
1faad4b0c1 add support for mistral json mode via anyscale 2023-12-29 22:26:22 +05:30
ishaan-jaff
7afc022ad3 (fix) counting streaming prompt tokens - azure 2023-12-29 16:13:52 +05:30
ishaan-jaff
4f832bce52 (fix) token_counter for tool calling 2023-12-29 15:54:03 +05:30
ishaan-jaff
806551ff99 (fix) use openai token counter for azure llms 2023-12-29 15:37:46 +05:30
ishaan-jaff
70376d3a4f (feat) azure stream - count correct prompt tokens 2023-12-29 15:15:39 +05:30
ishaan-jaff
8475fddc78 (feat) cloudflare - add exception mapping 2023-12-29 12:31:10 +05:30
ishaan-jaff
27f8598867 (feat) add cloudflare streaming 2023-12-29 12:01:26 +05:30
ishaan-jaff
c69f4f17a5 (feat) cloudflare - add optional params 2023-12-29 11:50:09 +05:30
ishaan-jaff
5d31bea9e0 (fix) tg AI cost tracking - zero-one-ai/Yi-34B-Chat 2023-12-29 09:14:07 +05:30
ishaan-jaff
362bed6ca3 (fix) together_ai cost tracking 2023-12-28 22:11:08 +05:30
Krrish Dholakia
5a48dac83f fix(vertex_ai.py): support function calling for gemini 2023-12-28 19:07:04 +05:30
ishaan-jaff
2a147579ec (feat) add voyage ai embeddings 2023-12-28 17:10:15 +05:30
Krrish Dholakia
507b6bf96e fix(utils.py): use local tiktoken copy 2023-12-28 11:22:33 +05:30
Krrish Dholakia
606de01ac0 fix(utils.py): allow text completion input to be either model or engine 2023-12-27 17:24:16 +05:30
ishaan-jaff
5f9e18c4c0 (fix) openai + stream - logprobs check 2023-12-27 16:59:56 +05:30
ishaan-jaff
c65d9a8b54 (feat) text-completion-openai, send 1 finish_reason 2023-12-27 15:45:40 +05:30
ishaan-jaff
592bcd5eea (fix) text_completion use correct finish reason 2023-12-27 15:20:26 +05:30
Krrish Dholakia
fd5e6efb1d fix(azure.py,-openai.py): correctly raise errors if streaming calls fail 2023-12-27 15:08:37 +05:30
Krrish Dholakia
85549c3d66 fix(google_kms.py): support enums for key management system 2023-12-27 13:19:33 +05:30
ishaan-jaff
021d7fab65 (feat) add text_completion, atext_completion CallTypes 2023-12-27 12:24:16 +05:30