Commit graph

2251 commits

Author SHA1 Message Date
Ishaan Jaff
ea3491a765 simplify calling azure/commmand-r-plus 2024-04-05 09:18:11 -07:00
Ishaan Jaff
442c053280 skip vertex test for this PR 2024-04-05 09:06:15 -07:00
Ishaan Jaff
12c5c96012 test - azure/command-r-plus 2024-04-05 08:56:05 -07:00
Ishaan Jaff
a0d429eadc (ci/cd) aws suspended account 2024-04-04 17:43:16 -07:00
Krrish Dholakia
9b8e1fe052 test(test_caching.py): skip test - aws suspended account
will need to recreate these objects on a new aws account
2024-04-04 15:07:19 -07:00
Krrish Dholakia
48a5948081 fix(router.py): handle id being passed in as int 2024-04-04 14:23:10 -07:00
Krrish Dholakia
a7da061ed8 test(test_router.py): fix casting 2024-04-04 13:54:16 -07:00
Krrish Dholakia
99e9b29526 test(test_router.py): fix test to check cast 2024-04-04 13:32:50 -07:00
Krrish Dholakia
0294c3f8a9 test(test_router.py): fix test to check type 2024-04-04 11:45:12 -07:00
Krrish Dholakia
f2823b6689 test(test_secret_manager.py): skip test - aws account suspended 2024-04-04 10:10:21 -07:00
Krrish Dholakia
c90da63650 test(test_completion.py): skip sagemaker test - aws account suspended 2024-04-04 09:52:24 -07:00
Krish Dholakia
a2dd945bd4 Merge pull request #2827 from BerriAI/litellm_model_add_api
fix(proxy_server.py): persist models added via `/model/new` to db
2024-04-03 23:30:39 -07:00
Krrish Dholakia
ec29f44c1f fix: raise correct error 2024-04-03 22:37:51 -07:00
Krrish Dholakia
a4a8129a13 fix(router.py): fix pydantic object logic 2024-04-03 21:57:19 -07:00
Krrish Dholakia
acdb3a7f7e test(test_acooldowns_router.py): fix tpm 2024-04-03 21:24:42 -07:00
Ishaan Jaff
bb9440fde6 (ci/cd) run again 2024-04-03 21:02:08 -07:00
Ishaan Jaff
91ec3478e1 test - voyage ai embedding 2024-04-03 20:54:35 -07:00
Krish Dholakia
b01e1c24df Merge branch 'main' into litellm_model_add_api 2024-04-03 20:29:44 -07:00
Krrish Dholakia
129bb52e9d fix(proxy_server.py): persist models added via /model/new to db
allows models to be used across instances

https://github.com/BerriAI/litellm/issues/2319 , https://github.com/BerriAI/litellm/issues/2329
2024-04-03 20:16:41 -07:00
Ishaan Jaff
0f6b31a49a ci/cd run again 2024-04-03 20:13:46 -07:00
Krrish Dholakia
7880dcbdf7 fix(openai.py): support passing prompt as list instead of concat string 2024-04-03 15:23:20 -07:00
Krrish Dholakia
f3fcfe3007 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
be3146c505 test(test_text_completion.py): unit testing for text completion pydantic object 2024-04-03 12:26:51 -07:00
Ishaan Jaff
13d0f7704b Merge pull request #2808 from BerriAI/litellm_use_all_proxy_team_models_auth
[feat] use `all-proxy-models` and `all-team-models` with Admin UI
2024-04-02 21:48:30 -07:00
Krrish Dholakia
21ade09a40 test(test_update_spend.py): fix test with right init 2024-04-02 21:11:26 -07:00
Ishaan Jaff
865cfd8c9f test new team request 2024-04-02 20:52:16 -07:00
Krrish Dholakia
029ee15951 perf(proxy_server.py): batch write spend logs
reduces prisma client errors, by batch writing spend logs - max 1k logs at a time
2024-04-02 18:46:55 -07:00
Ishaan Jaff
07885802d8 Merge pull request #2801 from BerriAI/litellm_support_all_models_as_a_ui_alias
[UI] use all_models alias
2024-04-02 17:53:25 -07:00
Ishaan Jaff
6534f887cc support all-proxy-models for teams 2024-04-02 16:04:09 -07:00
Ishaan Jaff
857a7c9135 support all-models-on-proxy 2024-04-02 15:52:54 -07:00
Ishaan Jaff
78edb29a29 (fix) support all-models alias on backend 2024-04-02 15:12:37 -07:00
Krrish Dholakia
ed46af19ec fix(openai.py): return logprobs for text completion calls 2024-04-02 14:05:56 -07:00
Krrish Dholakia
67f62aa53e fix(main.py): support text completion input being a list of strings
addresses - https://github.com/BerriAI/litellm/issues/2792, https://github.com/BerriAI/litellm/issues/2777
2024-04-02 08:50:16 -07:00
Ishaan Jaff
c2b9799e42 Merge pull request #2788 from BerriAI/litellm_support_-_models
[Feat] Allow using model = * on proxy config.yaml
2024-04-01 19:46:50 -07:00
Ishaan Jaff
71537393d3 test test_wildcard_openai_routing 2024-04-01 19:46:07 -07:00
Krrish Dholakia
e899e69a6a refactor: fix linting issue 2024-04-01 18:11:38 -07:00
Krrish Dholakia
e06d43dc90 fix(tpm_rpm_limiter.py): fix cache init logic 2024-04-01 18:01:38 -07:00
Krrish Dholakia
b2b8375987 fix(router.py): support context window fallbacks for pre-call checks 2024-04-01 10:51:54 -07:00
Krrish Dholakia
f68658b2da test(test_max_tpm_rpm_limiter.py): add unit testing for redis namespaces working for tpm/rpm limits 2024-04-01 10:39:03 -07:00
Krrish Dholakia
0a2a1c728b test(test_max_tpm_rpm_limiter.py): unit tests for key + team based tpm rpm limits on proxy 2024-04-01 08:11:30 -07:00
Ishaan Jaff
c6c6bd677e ci/cd run again 2024-04-01 07:40:05 -07:00
Krrish Dholakia
1a5409e52d test: fix test 2024-04-01 07:29:56 -07:00
Krrish Dholakia
9ec9936843 fix(utils.py): set redis_usage_cache to none by default 2024-04-01 07:29:56 -07:00
Krish Dholakia
1a30ad0d28 Merge pull request #2748 from BerriAI/litellm_anthropic_tool_calling_list_parsing_fix
fix(factory.py): parse list in xml tool calling response (anthropic)
2024-03-30 11:27:02 -07:00
Krrish Dholakia
7eff73c4e4 ci(config.yml): add lunary to circle ci 2024-03-29 22:09:21 -07:00
Vincelwt
35c279eff8 Merge branch 'main' into main 2024-03-30 13:21:53 +09:00
Krrish Dholakia
610b811033 test(test_key_generate_prisma.py): fix test 2024-03-29 20:30:43 -07:00
Krrish Dholakia
9fc8b8a215 fix(proxy_server.py): increment cached global proxy spend object 2024-03-29 20:02:31 -07:00
Krrish Dholakia
49e2624240 fix(proxy_server.py): enforce end user budgets with 'litellm.max_end_user_budget' param 2024-03-29 17:14:40 -07:00
Krrish Dholakia
5f6bb5c30a test(test_rules.py): fix assert 2024-03-29 13:12:16 -07:00