Commit graph

955 commits

Author SHA1 Message Date
ishaan-jaff
69552146f9 (feat) return custom_llm_provider in streaming response 2024-01-12 17:14:43 -08:00
Krrish Dholakia
813fb19620 fix: support async redis caching 2024-01-12 21:46:41 +05:30
David Leen
1660e4ab72 improve bedrock exception granularity 2024-01-12 16:38:55 +01:00
Krish Dholakia
7ecfc09221 Merge branch 'main' into litellm_embedding_caching_updates 2024-01-11 23:58:51 +05:30
Ishaan Jaff
8b0f93dd2d Merge pull request #1422 from dleen/httpx
(fix) create httpx.Request instead of httpx.request
2024-01-11 22:31:55 +05:30
David Leen
3cd5ef9a02 (fix) create httpx.Request instead of httpx.request
fixes #1420
2024-01-11 16:22:26 +01:00
Krrish Dholakia
107ba5aece fix(utils.py): use preset cache key for async calls as well 2024-01-11 19:30:28 +05:30
Krrish Dholakia
4ed9f2b712 fix(utils.py): bug fixes 2024-01-11 19:02:17 +05:30
Krrish Dholakia
63298d6c75 fix(utils.py): bug fixes 2024-01-11 18:14:22 +05:30
ishaan-jaff
50aeec537c (feat) improve litellm verbose logs 2024-01-11 18:13:08 +05:30
ishaan-jaff
859eb0d98e (feat) show args passed to litellm.completion, acompletion on call 2024-01-11 17:56:27 +05:30
Krrish Dholakia
66addb1a01 fix(utils.py): support caching individual items in embedding input list
https://github.com/BerriAI/litellm/issues/1350
2024-01-11 16:51:34 +05:30
Ishaan Jaff
6e1be43595 Merge pull request #1200 from MateoCamara/explicit-args-acomplete
feat: added explicit args to acomplete
2024-01-11 10:39:05 +05:30
ishaan-jaff
426ce91481 (fix) working s3 logging 2024-01-11 08:57:32 +05:30
Krrish Dholakia
3ed296e2dd fix(utils.py): raise correct error for azure content blocked error 2024-01-10 23:31:51 +05:30
Mateo Cámara
fb37ea291e Merge branch 'main' into explicit-args-acomplete 2024-01-09 13:07:37 +01:00
Ishaan Jaff
b09f38e835 Merge pull request #1381 from BerriAI/litellm_content_policy_violation_exception
[Feat] Add litellm.ContentPolicyViolationError
2024-01-09 17:18:29 +05:30
ishaan-jaff
650a6a8640 (chore) remove deprecated completion_with_config() tests 2024-01-09 17:13:06 +05:30
ishaan-jaff
1a4b522fbb (feat) add ContentPolicyViolationError for azure 2024-01-09 16:58:09 +05:30
ishaan-jaff
a6c45be9f2 (fix) ContentPolicyViolationError 2024-01-09 16:53:15 +05:30
Mateo Cámara
fb4b5eb4c1 Added test to check if acompletion is using the same parameters as CompletionRequest attributes. Added functools to client decorator to expose acompletion parameters from outside. 2024-01-09 12:06:49 +01:00
ishaan-jaff
66b23ecbb5 (v0) add ContentPolicyViolationError 2024-01-09 16:33:03 +05:30
ishaan-jaff
3081dc525a (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
5a7d2d9c80 fix(utils.py): error handling for litellm --model mistral edge case 2024-01-08 15:09:01 +05:30
Krrish Dholakia
5daa3ce237 fix(main.py): support cost calculation for text completion streaming object 2024-01-08 12:41:43 +05:30
Krrish Dholakia
f79113a7ac fix(utils.py): fix logging for text completion streaming 2024-01-08 12:05:28 +05:30
Krrish Dholakia
9cbb796874 fix(utils.py): fix exception raised 2024-01-08 07:42:17 +05:30
Krrish Dholakia
d8a2675643 fix(utils.py): map optional params for gemini 2024-01-08 07:38:55 +05:30
Krrish Dholakia
4c385a8cf5 bump: version 1.16.16 → 1.16.17 2024-01-08 07:16:37 +05:30
Krish Dholakia
67ecab4b38 Merge pull request #1344 from BerriAI/litellm_speed_improvements
Litellm speed improvements
2024-01-06 22:38:10 +05:30
Krrish Dholakia
2d1871a1ae fix(factory.py): support gemini-pro-vision on google ai studio
https://github.com/BerriAI/litellm/issues/1329
2024-01-06 22:36:22 +05:30
Krrish Dholakia
9c4cc06e64 fix(utils.py): handle original_response being a json 2024-01-06 17:02:50 +05:30
ishaan-jaff
eb8dba100d (fix) caching use same "created" in response_object 2024-01-05 16:03:56 +05:30
ishaan-jaff
9dc0074613 (feat) completion_cost: improve model=None error 2024-01-05 15:26:04 +05:30
ishaan-jaff
9313bda4c8 (feat) completion_cost - embeddings + raise Exception 2024-01-05 13:11:23 +05:30
Krrish Dholakia
52b78dc074 fix(utils.py): fix check for if cached response should be returned 2024-01-04 21:49:19 +05:30
Krrish Dholakia
6f7c764d5e fix(utils.py): raise a bad request error if litellm client raises a model /provider not found error 2024-01-04 15:50:43 +05:30
Krrish Dholakia
5aecbced99 feat(proxy_server.py): abstract config update/writing and support persisting config in db
allows user to opt into writing to db (SAVE_CONFIG_TO_DB) and removes any api keys before sending to db

 https://github.com/BerriAI/litellm/issues/1322
2024-01-04 14:44:59 +05:30
Krrish Dholakia
176af67aac fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
2024-01-03 12:42:43 +05:30
ishaan-jaff
f67b56ddf5 (fix) azure+stream: count completion tokens 2024-01-03 12:06:39 +05:30
ishaan-jaff
f582ef666f (fix) counting response tokens+streaming 2024-01-03 12:06:39 +05:30
Krrish Dholakia
adef0d9410 bump: version 1.16.10 → 1.16.11 2024-01-02 22:26:47 +05:30
ishaan-jaff
0e8809abf2 (feat) add xinference as an embedding provider 2024-01-02 15:32:26 +05:30
Krrish Dholakia
d4da63800e fix(utils.py): support token counting for gpt-4-vision models 2024-01-02 14:41:42 +05:30
Krrish Dholakia
4eae0c9a0d fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
2024-01-01 21:22:42 +05:30
ishaan-jaff
31bdcb48af (fix) use cloudflare optional params 2023-12-30 12:22:31 +05:30
Krrish Dholakia
7d55a563ee fix(main.py): don't set timeout as an optional api param 2023-12-30 11:47:07 +05:30
Krrish Dholakia
e1925d0e29 fix(router.py): support retry and fallbacks for atext_completion 2023-12-30 11:19:32 +05:30
ishaan-jaff
fa4a533e91 (fix) timeout optional param 2023-12-30 11:07:52 +05:30
ishaan-jaff
d5cbef4e36 (feat) proxy - support dynamic timeout per request 2023-12-30 10:55:42 +05:30