Commit graph

2623 commits

Author SHA1 Message Date
Ishaan Jaff
122dff007b fix gemini test 2024-04-05 20:38:28 -07:00
Ishaan Jaff
7fc416b636 (ci/cd) run again 2024-04-05 17:26:02 -07:00
Ishaan Jaff
fdadeabe79 fix testing yaml 2024-04-05 16:17:53 -07:00
Ishaan Jaff
faa0d38087
Merge pull request #2868 from BerriAI/litellm_add_command_r_on_proxy
Add Azure Command-r-plus on litellm proxy
2024-04-05 15:13:47 -07:00
Ishaan Jaff
2174b240d8
Merge pull request #2861 from BerriAI/litellm_add_azure_command_r_plust
[FEAT] add azure command-r-plus
2024-04-05 15:13:35 -07:00
Ishaan Jaff
9055a071e6 proxy - add azure/command r 2024-04-05 14:35:31 -07:00
Krrish Dholakia
9d7aaa91e0 test(test_amazing_vertex_completion.py): handle vertex ai rate limit errors 2024-04-05 14:04:29 -07:00
Krrish Dholakia
f7537f2cdf fix(test_amazing_vertex_completion.py): handle vertex ai rate limit error 2024-04-05 14:03:38 -07:00
Ishaan Jaff
3dbb547af0 catch gemini 429 errors 2024-04-05 14:02:01 -07:00
Ishaan Jaff
34a9b8648d fix gemini rate limit errors 2024-04-05 14:01:32 -07:00
Krrish Dholakia
f662dac5c9 fix(vertex_ai_anthropic.py): fix import 2024-04-05 13:38:05 -07:00
Ishaan Jaff
3b2c34d587 ci/cd - skip 429 errors from google gemini 2024-04-05 13:36:29 -07:00
Krrish Dholakia
695eb129ad fix(router.py): fix client init for streaming timeouts 2024-04-05 12:30:15 -07:00
Krrish Dholakia
f0c4ff6e60 fix(vertex_ai_anthropic.py): support streaming, async completion, async streaming for vertex ai anthropic 2024-04-05 09:27:48 -07:00
Ishaan Jaff
cfe358abaa simplify calling azure/commmand-r-plus 2024-04-05 09:18:11 -07:00
Ishaan Jaff
1ba1c84d83 skip vertex test for this PR 2024-04-05 09:06:15 -07:00
Ishaan Jaff
5d196ff300 test - azure/command-r-plus 2024-04-05 08:56:05 -07:00
Krish Dholakia
eb34306099
Merge pull request #2665 from BerriAI/litellm_claude_vertex_ai
[WIP] feat(vertex_ai_anthropic.py): Add support for claude 3 on vertex ai
2024-04-05 07:06:04 -07:00
Krish Dholakia
4ce8227e70
Merge pull request #2841 from Manouchehri/nuke-gemini-1.5-pro-vision
Fix: Remove non-existent gemini-1.5-pro-vision model.
2024-04-05 07:03:38 -07:00
Ishaan Jaff
7f50b80917 (ci/cd) aws suspended account 2024-04-04 17:43:16 -07:00
Krrish Dholakia
2472311a3f test(test_caching.py): skip test - aws suspended account
will need to recreate these objects on a new aws account
2024-04-04 15:07:19 -07:00
Krrish Dholakia
2236f283fe fix(router.py): handle id being passed in as int 2024-04-04 14:23:10 -07:00
Krrish Dholakia
2e40ab959d test(test_router.py): fix casting 2024-04-04 13:54:16 -07:00
Krrish Dholakia
c372c873a0 test(test_router.py): fix test to check cast 2024-04-04 13:32:50 -07:00
Krrish Dholakia
b9030be792 test(test_router.py): fix test to check type 2024-04-04 11:45:12 -07:00
David Manouchehri
6044045b91
Fix: Remove non-existent gemini-1.5-pro-vision model.
The gemini-1.5-pro model handles both text and vision.
2024-04-04 17:33:08 +00:00
Krrish Dholakia
c3e3ca9722 test(test_secret_manager.py): skip test - aws account suspended 2024-04-04 10:10:21 -07:00
Krrish Dholakia
dfcb6bcbc5 test(test_completion.py): skip sagemaker test - aws account suspended 2024-04-04 09:52:24 -07:00
Krish Dholakia
0c5b8a7667
Merge pull request #2827 from BerriAI/litellm_model_add_api
fix(proxy_server.py): persist models added via `/model/new` to db
2024-04-03 23:30:39 -07:00
Krrish Dholakia
346cd1876b fix: raise correct error 2024-04-03 22:37:51 -07:00
Krrish Dholakia
20849cbbfc fix(router.py): fix pydantic object logic 2024-04-03 21:57:19 -07:00
Krrish Dholakia
ef2f6ef6a2 test(test_acooldowns_router.py): fix tpm 2024-04-03 21:24:42 -07:00
Ishaan Jaff
fa44f45429 (ci/cd) run again 2024-04-03 21:02:08 -07:00
Ishaan Jaff
fb741d96ca test - voyage ai embedding 2024-04-03 20:54:35 -07:00
Krish Dholakia
6bc48d7e8d
Merge branch 'main' into litellm_model_add_api 2024-04-03 20:29:44 -07:00
Krrish Dholakia
f536fb13e6 fix(proxy_server.py): persist models added via /model/new to db
allows models to be used across instances

https://github.com/BerriAI/litellm/issues/2319 , https://github.com/BerriAI/litellm/issues/2329
2024-04-03 20:16:41 -07:00
Ishaan Jaff
d627c90bfd ci/cd run again 2024-04-03 20:13:46 -07:00
Krrish Dholakia
475144e5b7 fix(openai.py): support passing prompt as list instead of concat string 2024-04-03 15:23:20 -07:00
Krrish Dholakia
15e0099948 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
f17dd68df3 test(test_text_completion.py): unit testing for text completion pydantic object 2024-04-03 12:26:51 -07:00
Krrish Dholakia
1d341970ba feat(vertex_ai_anthropic.py): add claude 3 on vertex ai support - working .completions call
.completions() call works
2024-04-02 22:07:39 -07:00
Ishaan Jaff
4d76ec43ac
Merge pull request #2808 from BerriAI/litellm_use_all_proxy_team_models_auth
[feat] use `all-proxy-models` and `all-team-models` with Admin UI
2024-04-02 21:48:30 -07:00
Krrish Dholakia
b5ca4cc235 test(test_update_spend.py): fix test with right init 2024-04-02 21:11:26 -07:00
Ishaan Jaff
afd81f1609 test new team request 2024-04-02 20:52:16 -07:00
CLARKBENHAM
44cb0f352a formating 2024-04-02 19:56:07 -07:00
CLARKBENHAM
164898a213 fix lowest latency tests 2024-04-02 19:10:40 -07:00
CLARKBENHAM
29573b0967 param both tests to include failure (also fix prev) 2024-04-02 18:53:42 -07:00
Krrish Dholakia
d7601a4844 perf(proxy_server.py): batch write spend logs
reduces prisma client errors, by batch writing spend logs - max 1k logs at a time
2024-04-02 18:46:55 -07:00
CLARKBENHAM
4f95966475 tests showing error 2024-04-02 18:45:05 -07:00
Ishaan Jaff
21379eb56d
Merge pull request #2801 from BerriAI/litellm_support_all_models_as_a_ui_alias
[UI] use all_models alias
2024-04-02 17:53:25 -07:00