Commit graph

979 commits

Author SHA1 Message Date
Ishaan Jaff
62c83d36a5 (fix) remove litellm.telemetry 2024-03-26 11:21:09 -07:00
Krish Dholakia
f8d1d41b34 Merge pull request #2687 from BerriAI/litellm_jwt_auth_fixes_2
Litellm jwt auth fixes
2024-03-25 13:27:19 -07:00
Krrish Dholakia
38bdf31652 fix(handle_jwt.py): allow setting proxy admin role string for jwt auth 2024-03-25 12:20:14 -07:00
Ishaan Jaff
a6065e4853 (fix) stop using f strings in verbose logger 2024-03-25 10:55:30 -07:00
Ishaan Jaff
f0992c2dbd (fix) stop using f strings with logger 2024-03-25 10:47:18 -07:00
Ishaan Jaff
2c01457a4b (feat) stop eagerly evaluating fstring 2024-03-25 09:01:42 -07:00
Krrish Dholakia
ed3681d9bd fix(proxy_server.py): fix model info check 2024-03-23 15:59:17 -07:00
Krrish Dholakia
d275c1183e fix(proxy_server.py): enable jwt-auth for users
allow a user to auth into the proxy via jwt's and call allowed routes
2024-03-22 17:08:10 -07:00
Krrish Dholakia
f535a4498e fix(proxy_server.py): allow user to disable swagger ui docs via env
user can disable swagger ui docs by setting 'NO_DOCS="True"' in their env
2024-03-21 17:15:18 -07:00
Krish Dholakia
2009133d87 Merge branch 'main' into litellm_llm_api_prompt_injection_check 2024-03-21 09:57:10 -07:00
Ishaan Jaff
195783ff07 Merge pull request #2563 from eltociear/patch-2
Update proxy_server.py
2024-03-21 07:29:33 -07:00
Krrish Dholakia
e9cc6b4cc9 feat(proxy_server.py): enable llm api based prompt injection checks
run user calls through an llm api to check for prompt injection attacks. This happens in parallel to th
e actual llm call using `async_moderation_hook`
2024-03-20 22:43:42 -07:00
Krish Dholakia
e7ff074eab Merge pull request #2606 from BerriAI/litellm_jwt_auth_updates
fix(handle_jwt.py): track spend for user using jwt auth
2024-03-20 19:40:17 -07:00
Krrish Dholakia
feb78b7819 fix(proxy_server.py): fix import 2024-03-20 19:15:06 -07:00
Krrish Dholakia
f0d8472bfd fix(caching.py): enable async setting of cache for dual cache 2024-03-20 18:42:34 -07:00
Krrish Dholakia
9a221851d5 fix(handle_jwt.py): track spend for user using jwt auth 2024-03-20 10:55:52 -07:00
Ishaan Jaff
4ad5ff0230 (feat) better debugging for /cache/ping 2024-03-20 08:30:11 -07:00
Ishaan Jaff
e8f775ee04 (feat) litellm cache ping 2024-03-20 08:24:13 -07:00
Ishaan Jaff
4c38f5a450 (fix) caching - don't require cache password 2024-03-19 20:50:16 -07:00
Krrish Dholakia
5c88a8bcba fix(proxy_server.py): allow user to disable scheduled reset budget task 2024-03-19 20:36:22 -07:00
Krrish Dholakia
499087301c Revert "Merge pull request #2593 from BerriAI/litellm_reset_budget_fix"
This reverts commit afd363129f, reversing
changes made to c94bc94ad5.
2024-03-19 20:25:41 -07:00
Krish Dholakia
eeb8b585dc Merge pull request #2593 from BerriAI/litellm_reset_budget_fix
fix(proxy/utils.py): fix reset budget logic
2024-03-19 20:17:03 -07:00
Krrish Dholakia
5a49360af2 fix(proxy_server.py): add more debug logs 2024-03-19 19:59:43 -07:00
Krrish Dholakia
809124a698 fix: better debug logs 2024-03-19 19:28:26 -07:00
Ishaan Jaff
854437df0e Merge pull request #2591 from BerriAI/litellm_metrics_endpoint
[Feat] /metrics endpoint for Prometheus, Grafana
2024-03-19 18:08:22 -07:00
Ishaan Jaff
acd6691d6d (feat) using prom litellm 2024-03-19 15:49:23 -07:00
Krrish Dholakia
386688e5f9 feat(handle_jwt.py): support authenticating admins into the proxy via jwt's 2024-03-19 15:00:27 -07:00
Ishaan Jaff
3f28e932d9 (v0) prometheus metric 2024-03-19 14:48:38 -07:00
Krrish Dholakia
53ed290ed3 fix(proxy_server.py): expose disable_spend_logs flag in config general settings
Writing each spend log adds +300ms latency

https://github.com/BerriAI/litellm/issues/1714#issuecomment-1924727281
2024-03-19 12:08:37 -07:00
Krish Dholakia
f522a5236b Merge pull request #2561 from BerriAI/litellm_batch_writing_db
fix(proxy/utils.py): move to batch writing db updates
2024-03-18 21:50:47 -07:00
Krrish Dholakia
e20ee870d2 refactor(proxy_server.py): re-add custom db client logic - prevent regressions 2024-03-18 21:16:28 -07:00
Ishaan Jaff
1a3a19e195 (fix) if litellm-proxy-budget set use it 2024-03-18 20:31:23 -07:00
Krrish Dholakia
6c578db2c2 fix(proxy_server.py): fix spend log update 2024-03-18 20:26:28 -07:00
Ishaan Jaff
d2f3813e54 (fix) show global spend on UI 2024-03-18 18:15:08 -07:00
Krrish Dholakia
1b10123f07 fix(proxy/utils.py): batch writing updates to db 2024-03-18 16:47:02 -07:00
Krrish Dholakia
7f74cf0057 fix(proxy_server.py): don't override cache params on proxy config if set 2024-03-18 12:11:30 -07:00
Krrish Dholakia
23d088a1da fix(proxy_server.py): ignore cache if value is false 2024-03-18 11:15:32 -07:00
Ikko Eltociear Ashimine
c2adf3c247 Update proxy_server.py
intialize -> initialize
2024-03-18 01:28:12 +09:00
Krrish Dholakia
1618751824 fix(proxy/utils.py): move to batch writing db updates 2024-03-16 22:32:00 -07:00
Ishaan Jaff
0da6e78cb7 Merge pull request #2560 from BerriAI/litellm_view_team_based_spend
Admin UI - view team based spend
2024-03-16 19:38:17 -07:00
Ishaan Jaff
e6bebfda8f (feat) new ui build 2024-03-16 19:35:00 -07:00
Ishaan Jaff
12c68060c9 (feat) view team based spend 2024-03-16 19:06:16 -07:00
Krish Dholakia
293c32204e Merge pull request #2556 from BerriAI/litellm_aws_secret_manager_support
fix(utils.py): initial commit for aws secret manager support
2024-03-16 18:41:58 -07:00
Krrish Dholakia
bad2327b88 fix(utils.py): fix aws secret manager + support key_management_settings
fixes the aws secret manager implementation and allows the user to set which keys they want to check thr
ough it
2024-03-16 16:47:50 -07:00
Ishaan Jaff
3cc579dcbb Merge pull request #2559 from BerriAI/litellm_show_spend_correctly
(fix) admin ui - order spend by date
2024-03-16 16:43:13 -07:00
Ishaan Jaff
6253b45517 Merge pull request #2557 from BerriAI/litellm_clean_up_health_readiness
(fix) /health/readiness return success callback names as (str)
2024-03-16 16:38:01 -07:00
Ishaan Jaff
d6d9d69618 (fix) admin ui - order spend by date 2024-03-16 16:15:10 -07:00
Ishaan Jaff
e11e42a0ec (fix) /health readiness return callback names 2024-03-16 15:56:07 -07:00
Ishaan Jaff
dd17c5bd1a Merge pull request #2553 from BerriAI/litellm_improve_health_readiness_perf
(feat) Proxy - improve health readiness perf (93% faster)
2024-03-16 15:38:45 -07:00
Ishaan Jaff
e71a5b4d24 (fix) improve /health/readiness under high load 2024-03-16 14:14:17 -07:00