Commit graph

788 commits

Author SHA1 Message Date
Ishaan Jaff
b09f38e835 Merge pull request #1381 from BerriAI/litellm_content_policy_violation_exception
[Feat] Add litellm.ContentPolicyViolationError
2024-01-09 17:18:29 +05:30
ishaan-jaff
650a6a8640 (chore) remove deprecated completion_with_config() tests 2024-01-09 17:13:06 +05:30
ishaan-jaff
1a4b522fbb (feat) add ContentPolicyViolationError for azure 2024-01-09 16:58:09 +05:30
ishaan-jaff
a6c45be9f2 (fix) ContentPolicyViolationError 2024-01-09 16:53:15 +05:30
ishaan-jaff
66b23ecbb5 (v0) add ContentPolicyViolationError 2024-01-09 16:33:03 +05:30
ishaan-jaff
3081dc525a (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
5a7d2d9c80 fix(utils.py): error handling for litellm --model mistral edge case 2024-01-08 15:09:01 +05:30
Krrish Dholakia
5daa3ce237 fix(main.py): support cost calculation for text completion streaming object 2024-01-08 12:41:43 +05:30
Krrish Dholakia
f79113a7ac fix(utils.py): fix logging for text completion streaming 2024-01-08 12:05:28 +05:30
Krrish Dholakia
9cbb796874 fix(utils.py): fix exception raised 2024-01-08 07:42:17 +05:30
Krrish Dholakia
d8a2675643 fix(utils.py): map optional params for gemini 2024-01-08 07:38:55 +05:30
Krrish Dholakia
4c385a8cf5 bump: version 1.16.16 → 1.16.17 2024-01-08 07:16:37 +05:30
Krish Dholakia
67ecab4b38 Merge pull request #1344 from BerriAI/litellm_speed_improvements
Litellm speed improvements
2024-01-06 22:38:10 +05:30
Krrish Dholakia
2d1871a1ae fix(factory.py): support gemini-pro-vision on google ai studio
https://github.com/BerriAI/litellm/issues/1329
2024-01-06 22:36:22 +05:30
Krrish Dholakia
9c4cc06e64 fix(utils.py): handle original_response being a json 2024-01-06 17:02:50 +05:30
ishaan-jaff
eb8dba100d (fix) caching use same "created" in response_object 2024-01-05 16:03:56 +05:30
ishaan-jaff
9dc0074613 (feat) completion_cost: improve model=None error 2024-01-05 15:26:04 +05:30
ishaan-jaff
9313bda4c8 (feat) completion_cost - embeddings + raise Exception 2024-01-05 13:11:23 +05:30
Krrish Dholakia
52b78dc074 fix(utils.py): fix check for if cached response should be returned 2024-01-04 21:49:19 +05:30
Krrish Dholakia
6f7c764d5e fix(utils.py): raise a bad request error if litellm client raises a model /provider not found error 2024-01-04 15:50:43 +05:30
Krrish Dholakia
5aecbced99 feat(proxy_server.py): abstract config update/writing and support persisting config in db
allows user to opt into writing to db (SAVE_CONFIG_TO_DB) and removes any api keys before sending to db

 https://github.com/BerriAI/litellm/issues/1322
2024-01-04 14:44:59 +05:30
Krrish Dholakia
176af67aac fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
2024-01-03 12:42:43 +05:30
ishaan-jaff
f67b56ddf5 (fix) azure+stream: count completion tokens 2024-01-03 12:06:39 +05:30
ishaan-jaff
f582ef666f (fix) counting response tokens+streaming 2024-01-03 12:06:39 +05:30
Krrish Dholakia
adef0d9410 bump: version 1.16.10 → 1.16.11 2024-01-02 22:26:47 +05:30
ishaan-jaff
0e8809abf2 (feat) add xinference as an embedding provider 2024-01-02 15:32:26 +05:30
Krrish Dholakia
d4da63800e fix(utils.py): support token counting for gpt-4-vision models 2024-01-02 14:41:42 +05:30
Krrish Dholakia
4eae0c9a0d fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
2024-01-01 21:22:42 +05:30
ishaan-jaff
31bdcb48af (fix) use cloudflare optional params 2023-12-30 12:22:31 +05:30
Krrish Dholakia
7d55a563ee fix(main.py): don't set timeout as an optional api param 2023-12-30 11:47:07 +05:30
Krrish Dholakia
e1925d0e29 fix(router.py): support retry and fallbacks for atext_completion 2023-12-30 11:19:32 +05:30
ishaan-jaff
fa4a533e91 (fix) timeout optional param 2023-12-30 11:07:52 +05:30
ishaan-jaff
d5cbef4e36 (feat) proxy - support dynamic timeout per request 2023-12-30 10:55:42 +05:30
Marmik Pandya
1faad4b0c1 add support for mistral json mode via anyscale 2023-12-29 22:26:22 +05:30
ishaan-jaff
7afc022ad3 (fix) counting streaming prompt tokens - azure 2023-12-29 16:13:52 +05:30
ishaan-jaff
4f832bce52 (fix) token_counter for tool calling 2023-12-29 15:54:03 +05:30
ishaan-jaff
806551ff99 (fix) use openai token counter for azure llms 2023-12-29 15:37:46 +05:30
ishaan-jaff
70376d3a4f (feat) azure stream - count correct prompt tokens 2023-12-29 15:15:39 +05:30
ishaan-jaff
8475fddc78 (feat) cloudflare - add exception mapping 2023-12-29 12:31:10 +05:30
ishaan-jaff
27f8598867 (feat) add cloudflare streaming 2023-12-29 12:01:26 +05:30
ishaan-jaff
c69f4f17a5 (feat) cloudflare - add optional params 2023-12-29 11:50:09 +05:30
ishaan-jaff
5d31bea9e0 (fix) tg AI cost tracking - zero-one-ai/Yi-34B-Chat 2023-12-29 09:14:07 +05:30
ishaan-jaff
362bed6ca3 (fix) together_ai cost tracking 2023-12-28 22:11:08 +05:30
Krrish Dholakia
5a48dac83f fix(vertex_ai.py): support function calling for gemini 2023-12-28 19:07:04 +05:30
ishaan-jaff
2a147579ec (feat) add voyage ai embeddings 2023-12-28 17:10:15 +05:30
Krrish Dholakia
507b6bf96e fix(utils.py): use local tiktoken copy 2023-12-28 11:22:33 +05:30
Krrish Dholakia
606de01ac0 fix(utils.py): allow text completion input to be either model or engine 2023-12-27 17:24:16 +05:30
ishaan-jaff
5f9e18c4c0 (fix) openai + stream - logprobs check 2023-12-27 16:59:56 +05:30
ishaan-jaff
c65d9a8b54 (feat) text-completion-openai, send 1 finish_reason 2023-12-27 15:45:40 +05:30
ishaan-jaff
592bcd5eea (fix) text_completion use correct finish reason 2023-12-27 15:20:26 +05:30