Commit graph

1641 commits

Author SHA1 Message Date
Krish Dholakia
b01e1c24df Merge branch 'main' into litellm_model_add_api 2024-04-03 20:29:44 -07:00
Krrish Dholakia
db9f674937 build(schema.prisma): add proxy model table to prisma schema 2024-04-03 20:17:38 -07:00
Krrish Dholakia
129bb52e9d fix(proxy_server.py): persist models added via /model/new to db
allows models to be used across instances

https://github.com/BerriAI/litellm/issues/2319 , https://github.com/BerriAI/litellm/issues/2329
2024-04-03 20:16:41 -07:00
Ishaan Jaff
984a08a500 ui - new build 2024-04-03 19:42:26 -07:00
Ishaan Jaff
13c1130d90 fix team update bug 2024-04-03 19:38:07 -07:00
Krrish Dholakia
5fe8aa27d1 feat(prometheus.py): track team based metrics on prometheus 2024-04-03 13:43:21 -07:00
Krrish Dholakia
f3fcfe3007 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
2d0c450079 fix(proxy_cli.py): revert db timeout change - user-controllable param
db timeout is a user controllable param, not necessary to change defaults
2024-04-03 09:37:57 -07:00
Krrish Dholakia
895073e68a fix(proxy_server.py): bump default db timeouts 2024-04-03 09:35:08 -07:00
Krrish Dholakia
96e0078b3a fix(_types.py): make /metrics public 2024-04-03 09:17:40 -07:00
Ishaan Jaff
8117fd6364 Merge pull request #2816 from BerriAI/litellm_ui_fix_viewing_spend
UI Bug fix - Unable to generate keys / start UI
2024-04-03 07:57:58 -07:00
Krrish Dholakia
71b8a60d14 fix(proxy_server.py): support calling public endpoints when jwt_auth is enabled 2024-04-03 07:56:53 -07:00
Ishaan Jaff
f990984bf0 new ui build 2024-04-03 07:56:29 -07:00
Ishaan Jaff
b47de20862 new ui build 2024-04-02 21:49:36 -07:00
Ishaan Jaff
a46179947e fix safe use token id 2024-04-02 21:40:35 -07:00
Ishaan Jaff
734c4fba3b v0 use token_in /key_generate 2024-04-02 21:31:24 -07:00
Ishaan Jaff
ec1ba2e216 proxy test all-tea-models 2024-04-02 20:50:47 -07:00
Krrish Dholakia
029ee15951 perf(proxy_server.py): batch write spend logs
reduces prisma client errors, by batch writing spend logs - max 1k logs at a time
2024-04-02 18:46:55 -07:00
Ishaan Jaff
6534f887cc support all-proxy-models for teams 2024-04-02 16:04:09 -07:00
Ishaan Jaff
857a7c9135 support all-models-on-proxy 2024-04-02 15:52:54 -07:00
Ishaan Jaff
78edb29a29 (fix) support all-models alias on backend 2024-04-02 15:12:37 -07:00
Ishaan Jaff
60e449f03c fix show correct team based usage 2024-04-02 13:43:33 -07:00
Ishaan Jaff
07821e213a fix left join on litellm team table 2024-04-02 13:36:22 -07:00
Ishaan Jaff
dc4741a58b Merge pull request #2795 from BerriAI/ui_edit_keys
[UI] Edit Keys
2024-04-02 12:59:01 -07:00
Ishaan Jaff
119af72183 new ui build 2024-04-02 12:58:31 -07:00
Krish Dholakia
6a1e08125c Merge pull request #2789 from BerriAI/litellm_set_ttl
fix(proxy_server.py): allow user to set in-memory + redis ttl
2024-04-02 08:53:34 -07:00
Ishaan Jaff
c2b9799e42 Merge pull request #2788 from BerriAI/litellm_support_-_models
[Feat] Allow using model = * on proxy config.yaml
2024-04-01 19:46:50 -07:00
Krish Dholakia
6d3bee777e Merge pull request #2787 from BerriAI/litellm_optional_team_jwt_claim
fix(proxy_server.py): don't require scope for team-based jwt access
2024-04-01 19:16:39 -07:00
Krrish Dholakia
74a943bc62 fix(proxy_server.py): fix cache param arg name 2024-04-01 19:14:39 -07:00
Krrish Dholakia
aa5f88ec45 fix(proxy_server.py): allow user to set in-memory + redis ttl
addresses - https://github.com/BerriAI/litellm/issues/2700
2024-04-01 19:13:23 -07:00
Ishaan Jaff
716fcd3ec4 (fix) allow wildcard models 2024-04-01 19:07:05 -07:00
Krrish Dholakia
abf87fe448 fix(proxy_server.py): don't require scope for team-based jwt access
If team with the client_id exists then it should be allowed to make a request, if it doesn't then as we discussed it should return an error
2024-04-01 18:52:00 -07:00
Ishaan Jaff
96d7377824 Merge pull request #2785 from BerriAI/litellm_high_traffic_redis_caching_fixes
[Feat] Proxy - high traffic redis caching - when using `url`
2024-04-01 18:38:27 -07:00
Krrish Dholakia
e06d43dc90 fix(tpm_rpm_limiter.py): fix cache init logic 2024-04-01 18:01:38 -07:00
Ishaan Jaff
86ba03e08a add /cache/redis/info endpoint 2024-04-01 16:51:23 -07:00
Krrish Dholakia
8d35e659ad fix(proxy/utils.py): support redis caching for alerting 2024-04-01 16:13:59 -07:00
Krrish Dholakia
91120df3f7 docs(openai.md): fix docs to include example of calling openai on proxy 2024-04-01 12:09:22 -07:00
Ishaan Jaff
aff63e556c (fix) _update_end_user_cache 2024-04-01 11:18:00 -07:00
Krrish Dholakia
f68658b2da test(test_max_tpm_rpm_limiter.py): add unit testing for redis namespaces working for tpm/rpm limits 2024-04-01 10:39:03 -07:00
Krrish Dholakia
b39bc583bd test(test_max_tpm_rpm_limiter.py): unit tests for key + team based tpm rpm limits on proxy 2024-04-01 08:00:01 -07:00
Krish Dholakia
f635d15c2b Merge pull request #2775 from BerriAI/litellm_redis_user_api_key_cache_v3
fix(tpm_rpm_limiter.py): enable redis caching for tpm/rpm checks on keys/user/teams
2024-03-30 22:07:05 -07:00
Krrish Dholakia
60b9e25e3c fix(proxy/utils.py): uncomment max parallel request limit check 2024-03-30 20:51:59 -07:00
Krrish Dholakia
7738107d49 fix(utils.py): set redis_usage_cache to none by default 2024-03-30 20:10:56 -07:00
Krrish Dholakia
555f0af027 fix(tpm_rpm_limiter.py): enable redis caching for tpm/rpm checks on keys/user/teams
allows tpm/rpm checks to work across instances

https://github.com/BerriAI/litellm/issues/2730
2024-03-30 20:01:36 -07:00
Ishaan Jaff
5577c4b821 (ui) new ui build 2024-03-30 14:38:19 -07:00
Ishaan Jaff
e1f3634855 (ui) show proxy spend 2024-03-30 14:02:43 -07:00
Ishaan Jaff
9e0e389109 (ui) view spend by team name on usage 2024-03-30 13:25:32 -07:00
Ishaan Jaff
bad8af69f2 (ui) new build 2024-03-30 12:17:07 -07:00
Ishaan Jaff
6024ae851d Merge branch 'main' into litellm_edit_teams 2024-03-30 12:15:53 -07:00
Ishaan Jaff
2fade23b27 (new) ui build 2024-03-30 12:14:43 -07:00