Commit graph

4168 commits

Author SHA1 Message Date
ishaan-jaff
8197b3de0a (fix) remove extra statement 2024-02-07 19:24:27 -08:00
ishaan-jaff
8425a8ba22 (fix) track cost for semantic_caching, place on langfuse trace 2024-02-07 19:21:50 -08:00
Krrish Dholakia
d2dceb3537 fix(proxy_server.py): check if prisma client is set before scheduling reset budget 2024-02-07 18:14:37 -08:00
ishaan-jaff
68926c6524 (fix) model_prices_and_context_window.json error 2024-02-07 15:42:37 -08:00
Krrish Dholakia
655fcd4d79 fix(utils.py): fix ollama stop sequence mapping 2024-02-07 13:14:03 -08:00
ishaan-jaff
258fe63e7d (fix) ui - when request body is None 2024-02-07 11:33:43 -08:00
Krrish Dholakia
8939593826 fix(proxy_server.py): fix merge errors 2024-02-07 00:04:52 -08:00
Krrish Dholakia
184e78772b refactor(proxy_server.py): fix merge error 2024-02-06 23:44:23 -08:00
Krrish Dholakia
46dd08c207 refactor(main.py): trigger rebuild 2024-02-06 23:39:28 -08:00
Krish Dholakia
df60edfa07
Merge branch 'main' into litellm_spend_logging_high_traffic 2024-02-06 23:36:58 -08:00
Krrish Dholakia
fd9c7a90af fix(proxy_server.py): update user cache to with new spend 2024-02-06 23:06:05 -08:00
Krrish Dholakia
73d8e3e640 fix(ollama_chat.py): fix token counting 2024-02-06 22:18:46 -08:00
Krrish Dholakia
4174471dac fix(proxy_server.py): fix endpoint 2024-02-06 22:09:30 -08:00
Krish Dholakia
2bc710d8e9
Merge pull request #1843 from BerriAI/litellm_admin_ui_view_all_keys
feat(ui): enable admin to view all valid keys created on the proxy
2024-02-06 22:06:46 -08:00
Krrish Dholakia
0874c17a31 fix: export npm build into proxy 2024-02-06 20:12:50 -08:00
Krrish Dholakia
4a0df3cb4f fix(proxy_cli.py-&&-proxy_server.py): bump reset budget intervals and fix pool limits for prisma connections 2024-02-06 19:39:49 -08:00
ishaan-jaff
5a29f362ee (fix) allow litellm_settings to be None 2024-02-06 19:29:39 -08:00
ishaan-jaff
7b26b3b789 (ci/cd) run again 2024-02-06 18:25:15 -08:00
Krrish Dholakia
b6adeec347 fix(proxy_server.py): prisma client fixes for high traffic 2024-02-06 17:30:36 -08:00
Ishaan Jaff
73c6ce890b
Merge pull request #1859 from BerriAI/litellm_allow_using_budgets_without_keys
[Feat] Budgets for 'user' param passed to /chat/completions, /embeddings etc
2024-02-06 16:32:25 -08:00
ishaan-jaff
6369424629 (ci/cd) run again 2024-02-06 16:08:25 -08:00
Krish Dholakia
9e9fb747ce
Merge branch 'main' into litellm_slack_langfuse_alerting 2024-02-06 15:48:52 -08:00
ishaan-jaff
196787359f (test) track_cost_ for end users 2024-02-06 15:25:51 -08:00
ishaan-jaff
52b864976b (feat) support max_user_budget 2024-02-06 15:19:36 -08:00
Krrish Dholakia
be81183782 refactor(main.py): trigger deploy
n
2024-02-06 15:17:40 -08:00
ishaan-jaff
8ba2c8dbf7 (fix) langfuse show semantic-similarity in tags 2024-02-06 15:17:40 -08:00
ishaan-jaff
325ca43946 (feat) show semantic-cache on health/readiness 2024-02-06 15:17:40 -08:00
Krrish Dholakia
0d03b28a3b test(test_completion.py): fix test 2024-02-06 15:17:40 -08:00
ishaan-jaff
b5db630dba (ci/cd) run again 2024-02-06 15:17:40 -08:00
ishaan-jaff
43061d612d (fix) mark semantic caching as beta test 2024-02-06 15:17:40 -08:00
ishaan-jaff
e32c2beddd (fix) semantic caching 2024-02-06 15:17:40 -08:00
ishaan-jaff
f3de05cc54 (fix) test-semantic caching 2024-02-06 15:17:40 -08:00
ishaan-jaff
f8248b2c79 (feat) redis-semantic cache on proxy 2024-02-06 15:17:40 -08:00
ishaan-jaff
58f47c9e29 (fix) use semantic cache on proxy 2024-02-06 15:17:40 -08:00
ishaan-jaff
91424b66d7 allow setting redis_semantic cache_embedding model 2024-02-06 15:17:40 -08:00
ishaan-jaff
e2c88ce154 (feat) log semantic_sim to langfuse 2024-02-06 15:17:40 -08:00
ishaan-jaff
e0d5c953d6 (feat) working semantic cache on proxy 2024-02-06 15:17:40 -08:00
ishaan-jaff
c8d5714e59 (feat) redis-semantic cache 2024-02-06 15:17:35 -08:00
ishaan-jaff
aa7580411d (feat) working semantic-cache on litellm proxy 2024-02-06 15:17:35 -08:00
ishaan-jaff
1d151e4777 (test) async semantic cache 2024-02-06 15:17:35 -08:00
ishaan-jaff
08d72fd2a0 (feat) RedisSemanticCache - async 2024-02-06 15:17:35 -08:00
ishaan-jaff
553b993473 (fix) semantic cache 2024-02-06 15:17:35 -08:00
ishaan-jaff
2ad8b70f50 (test) semantic caching 2024-02-06 15:17:35 -08:00
ishaan-jaff
a5afbf6d56 (test) semantic cache 2024-02-06 15:17:35 -08:00
ishaan-jaff
a510adb1e6 (feat) working - sync semantic caching 2024-02-06 15:17:35 -08:00
ishaan-jaff
d67a9ada4f (feat )add semantic cache 2024-02-06 15:17:35 -08:00
ishaan-jaff
d85b1f8816 (feat) show langfuse logging tags better through proxy 2024-02-06 15:17:35 -08:00
Krrish Dholakia
eee5353e77 fix(utils.py): round max tokens to be int always 2024-02-06 15:17:35 -08:00
ishaan-jaff
647dbb9331 (ci/cd) run again 2024-02-06 15:17:35 -08:00
ishaan-jaff
01701c95b8 (ci/cd) run again 2024-02-06 15:17:35 -08:00