Commit graph

1139 commits

Author SHA1 Message Date
Krrish Dholakia
e38ff06d01 fix(proxy_server.py): add more detailed debugging 2024-04-08 17:37:55 -07:00
Ishaan Jaff
2a69650e8c
Merge pull request #2903 from BerriAI/litellm_set_new_callbacks_ui
ui - add langfuse modal
2024-04-08 17:37:38 -07:00
Ishaan Jaff
90656bd203 ui - set langfuse callbacks 2024-04-08 17:32:20 -07:00
Krrish Dholakia
8f1872eaf3 fix(proxy_server.py): allow for no models in model_list - all models passed in via /model/new 2024-04-08 16:17:52 -07:00
Ishaan Jaff
0d925a6c55
Merge pull request #2902 from BerriAI/litellm_ui_set_get_callbacks
UI view set callbacks
2024-04-08 16:08:14 -07:00
Ishaan Jaff
3ee345477e proxy - add callbacks 2024-04-08 15:59:04 -07:00
Ishaan Jaff
de14e11d3e ui - return callbacks set on litellm 2024-04-08 15:25:11 -07:00
Krrish Dholakia
baed85c2c2 fix(proxy_server.py): add info statement 2024-04-08 15:12:56 -07:00
Krrish Dholakia
5e46432dfe docs(demo.md): fix iframe link 2024-04-08 15:11:19 -07:00
Ishaan Jaff
cb0350ef29 feat - write success callbacks to db 2024-04-08 15:03:18 -07:00
Krrish Dholakia
009f548079 fix(proxy_server.py): allow /model/new feature flag to work via env 2024-04-08 14:57:19 -07:00
Krrish Dholakia
6110d32b1c feat(proxy/utils.py): return api base for request hanging alerts 2024-04-06 15:58:53 -07:00
Krrish Dholakia
5d1a813b66 feat(prisma.schema): allow setting user_alias 2024-04-05 17:04:14 -07:00
Krrish Dholakia
7949f4849e fix(proxy_server.py): remove original response from headers - causing nginx to fail
response headers > 4kb cause nginx to fail
2024-04-05 12:36:14 -07:00
Krish Dholakia
24e2535441
Merge branch 'main' into litellm_add_model_api_fix 2024-04-04 19:43:52 -07:00
Krrish Dholakia
1ee78c39e9 fix(proxy_server.py): skip to next model if invalid model (raise logger.error) 2024-04-04 19:22:17 -07:00
Krrish Dholakia
fbfcd57798 fix(proxy_server.py): fix linting issue 2024-04-04 19:15:57 -07:00
Krrish Dholakia
e3c2bdef4d feat(ui): add models via ui
adds ability to add models via ui to the proxy. also fixes additional bugs around new /model/new endpoint
2024-04-04 18:56:20 -07:00
Ishaan Jaff
ac5507bd84 ui show spend per tag 2024-04-04 16:57:45 -07:00
Ishaan Jaff
1119cc49a8
Merge pull request #2840 from BerriAI/litellm_return_cache_key_responses
[FEAT] Proxy - Delete Cache Keys + return cache key in responses
2024-04-04 11:52:52 -07:00
Ishaan Jaff
7e1d5c81b4 return cache key in streming responses 2024-04-04 11:00:00 -07:00
Ishaan Jaff
c4cb0afa98 feat - delete cache key 2024-04-04 10:56:47 -07:00
Krrish Dholakia
592241b4eb fix(proxy_server.py): fix linting error 2024-04-04 10:40:32 -07:00
Ishaan Jaff
9dc4127576 v0 return cache key in responses 2024-04-04 10:11:18 -07:00
Krrish Dholakia
4b56f08cbe test(test_models.py): fix delete model test 2024-04-04 08:46:08 -07:00
Krrish Dholakia
346cd1876b fix: raise correct error 2024-04-03 22:37:51 -07:00
Krish Dholakia
6bc48d7e8d
Merge branch 'main' into litellm_model_add_api 2024-04-03 20:29:44 -07:00
Krrish Dholakia
f536fb13e6 fix(proxy_server.py): persist models added via /model/new to db
allows models to be used across instances

https://github.com/BerriAI/litellm/issues/2319 , https://github.com/BerriAI/litellm/issues/2329
2024-04-03 20:16:41 -07:00
Ishaan Jaff
6edaaa92ab fix team update bug 2024-04-03 19:38:07 -07:00
Krrish Dholakia
15e0099948 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
8f24202c83 fix(proxy_server.py): support calling public endpoints when jwt_auth is enabled 2024-04-03 07:56:53 -07:00
Ishaan Jaff
8a8233e428 fix safe use token id 2024-04-02 21:40:35 -07:00
Ishaan Jaff
15685a8f53 v0 use token_in /key_generate 2024-04-02 21:31:24 -07:00
Ishaan Jaff
1aeccf3f0a proxy test all-tea-models 2024-04-02 20:50:47 -07:00
Krrish Dholakia
d7601a4844 perf(proxy_server.py): batch write spend logs
reduces prisma client errors, by batch writing spend logs - max 1k logs at a time
2024-04-02 18:46:55 -07:00
Ishaan Jaff
3245d8cdce support all-proxy-models for teams 2024-04-02 16:04:09 -07:00
Ishaan Jaff
b83c452ddd support all-models-on-proxy 2024-04-02 15:52:54 -07:00
Ishaan Jaff
73ef4780f7 (fix) support all-models alias on backend 2024-04-02 15:12:37 -07:00
Ishaan Jaff
3d32567f4c fix show correct team based usage 2024-04-02 13:43:33 -07:00
Ishaan Jaff
327cf73d73 fix left join on litellm team table 2024-04-02 13:36:22 -07:00
Krish Dholakia
7233e5ab25
Merge pull request #2789 from BerriAI/litellm_set_ttl
fix(proxy_server.py): allow user to set in-memory + redis ttl
2024-04-02 08:53:34 -07:00
Ishaan Jaff
92984a1c6f
Merge pull request #2788 from BerriAI/litellm_support_-_models
[Feat] Allow using model = * on proxy config.yaml
2024-04-01 19:46:50 -07:00
Krish Dholakia
da85384649
Merge pull request #2787 from BerriAI/litellm_optional_team_jwt_claim
fix(proxy_server.py): don't require scope for team-based jwt access
2024-04-01 19:16:39 -07:00
Krrish Dholakia
c096ba566f fix(proxy_server.py): fix cache param arg name 2024-04-01 19:14:39 -07:00
Krrish Dholakia
203e2776f8 fix(proxy_server.py): allow user to set in-memory + redis ttl
addresses - https://github.com/BerriAI/litellm/issues/2700
2024-04-01 19:13:23 -07:00
Ishaan Jaff
037b624c89 (fix) allow wildcard models 2024-04-01 19:07:05 -07:00
Krrish Dholakia
c52819d47c fix(proxy_server.py): don't require scope for team-based jwt access
If team with the client_id exists then it should be allowed to make a request, if it doesn't then as we discussed it should return an error
2024-04-01 18:52:00 -07:00
Ishaan Jaff
b14b6083f5
Merge pull request #2785 from BerriAI/litellm_high_traffic_redis_caching_fixes
[Feat] Proxy - high traffic redis caching - when using `url`
2024-04-01 18:38:27 -07:00
Krrish Dholakia
6467dd4e11 fix(tpm_rpm_limiter.py): fix cache init logic 2024-04-01 18:01:38 -07:00
Ishaan Jaff
9accc544e9 add /cache/redis/info endpoint 2024-04-01 16:51:23 -07:00