Ishaan Jaff
|
122dff007b
|
fix gemini test
|
2024-04-05 20:38:28 -07:00 |
|
Ishaan Jaff
|
7fc416b636
|
(ci/cd) run again
|
2024-04-05 17:26:02 -07:00 |
|
Ishaan Jaff
|
fdadeabe79
|
fix testing yaml
|
2024-04-05 16:17:53 -07:00 |
|
Ishaan Jaff
|
faa0d38087
|
Merge pull request #2868 from BerriAI/litellm_add_command_r_on_proxy
Add Azure Command-r-plus on litellm proxy
|
2024-04-05 15:13:47 -07:00 |
|
Ishaan Jaff
|
2174b240d8
|
Merge pull request #2861 from BerriAI/litellm_add_azure_command_r_plust
[FEAT] add azure command-r-plus
|
2024-04-05 15:13:35 -07:00 |
|
Ishaan Jaff
|
9055a071e6
|
proxy - add azure/command r
|
2024-04-05 14:35:31 -07:00 |
|
Krrish Dholakia
|
9d7aaa91e0
|
test(test_amazing_vertex_completion.py): handle vertex ai rate limit errors
|
2024-04-05 14:04:29 -07:00 |
|
Krrish Dholakia
|
f7537f2cdf
|
fix(test_amazing_vertex_completion.py): handle vertex ai rate limit error
|
2024-04-05 14:03:38 -07:00 |
|
Ishaan Jaff
|
3dbb547af0
|
catch gemini 429 errors
|
2024-04-05 14:02:01 -07:00 |
|
Ishaan Jaff
|
34a9b8648d
|
fix gemini rate limit errors
|
2024-04-05 14:01:32 -07:00 |
|
Krrish Dholakia
|
f662dac5c9
|
fix(vertex_ai_anthropic.py): fix import
|
2024-04-05 13:38:05 -07:00 |
|
Ishaan Jaff
|
3b2c34d587
|
ci/cd - skip 429 errors from google gemini
|
2024-04-05 13:36:29 -07:00 |
|
Krrish Dholakia
|
695eb129ad
|
fix(router.py): fix client init for streaming timeouts
|
2024-04-05 12:30:15 -07:00 |
|
Krrish Dholakia
|
f0c4ff6e60
|
fix(vertex_ai_anthropic.py): support streaming, async completion, async streaming for vertex ai anthropic
|
2024-04-05 09:27:48 -07:00 |
|
Ishaan Jaff
|
cfe358abaa
|
simplify calling azure/commmand-r-plus
|
2024-04-05 09:18:11 -07:00 |
|
Ishaan Jaff
|
1ba1c84d83
|
skip vertex test for this PR
|
2024-04-05 09:06:15 -07:00 |
|
Ishaan Jaff
|
5d196ff300
|
test - azure/command-r-plus
|
2024-04-05 08:56:05 -07:00 |
|
Krish Dholakia
|
eb34306099
|
Merge pull request #2665 from BerriAI/litellm_claude_vertex_ai
[WIP] feat(vertex_ai_anthropic.py): Add support for claude 3 on vertex ai
|
2024-04-05 07:06:04 -07:00 |
|
Krish Dholakia
|
4ce8227e70
|
Merge pull request #2841 from Manouchehri/nuke-gemini-1.5-pro-vision
Fix: Remove non-existent gemini-1.5-pro-vision model.
|
2024-04-05 07:03:38 -07:00 |
|
Ishaan Jaff
|
7f50b80917
|
(ci/cd) aws suspended account
|
2024-04-04 17:43:16 -07:00 |
|
Krrish Dholakia
|
2472311a3f
|
test(test_caching.py): skip test - aws suspended account
will need to recreate these objects on a new aws account
|
2024-04-04 15:07:19 -07:00 |
|
Krrish Dholakia
|
2236f283fe
|
fix(router.py): handle id being passed in as int
|
2024-04-04 14:23:10 -07:00 |
|
Krrish Dholakia
|
2e40ab959d
|
test(test_router.py): fix casting
|
2024-04-04 13:54:16 -07:00 |
|
Krrish Dholakia
|
c372c873a0
|
test(test_router.py): fix test to check cast
|
2024-04-04 13:32:50 -07:00 |
|
Krrish Dholakia
|
b9030be792
|
test(test_router.py): fix test to check type
|
2024-04-04 11:45:12 -07:00 |
|
David Manouchehri
|
6044045b91
|
Fix: Remove non-existent gemini-1.5-pro-vision model.
The gemini-1.5-pro model handles both text and vision.
|
2024-04-04 17:33:08 +00:00 |
|
Krrish Dholakia
|
c3e3ca9722
|
test(test_secret_manager.py): skip test - aws account suspended
|
2024-04-04 10:10:21 -07:00 |
|
Krrish Dholakia
|
dfcb6bcbc5
|
test(test_completion.py): skip sagemaker test - aws account suspended
|
2024-04-04 09:52:24 -07:00 |
|
Krish Dholakia
|
0c5b8a7667
|
Merge pull request #2827 from BerriAI/litellm_model_add_api
fix(proxy_server.py): persist models added via `/model/new` to db
|
2024-04-03 23:30:39 -07:00 |
|
Krrish Dholakia
|
346cd1876b
|
fix: raise correct error
|
2024-04-03 22:37:51 -07:00 |
|
Krrish Dholakia
|
20849cbbfc
|
fix(router.py): fix pydantic object logic
|
2024-04-03 21:57:19 -07:00 |
|
Krrish Dholakia
|
ef2f6ef6a2
|
test(test_acooldowns_router.py): fix tpm
|
2024-04-03 21:24:42 -07:00 |
|
Ishaan Jaff
|
fa44f45429
|
(ci/cd) run again
|
2024-04-03 21:02:08 -07:00 |
|
Ishaan Jaff
|
fb741d96ca
|
test - voyage ai embedding
|
2024-04-03 20:54:35 -07:00 |
|
Krish Dholakia
|
6bc48d7e8d
|
Merge branch 'main' into litellm_model_add_api
|
2024-04-03 20:29:44 -07:00 |
|
Krrish Dholakia
|
f536fb13e6
|
fix(proxy_server.py): persist models added via /model/new to db
allows models to be used across instances
https://github.com/BerriAI/litellm/issues/2319 , https://github.com/BerriAI/litellm/issues/2329
|
2024-04-03 20:16:41 -07:00 |
|
Ishaan Jaff
|
d627c90bfd
|
ci/cd run again
|
2024-04-03 20:13:46 -07:00 |
|
Krrish Dholakia
|
475144e5b7
|
fix(openai.py): support passing prompt as list instead of concat string
|
2024-04-03 15:23:20 -07:00 |
|
Krrish Dholakia
|
15e0099948
|
fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
|
2024-04-03 13:05:43 -07:00 |
|
Krrish Dholakia
|
f17dd68df3
|
test(test_text_completion.py): unit testing for text completion pydantic object
|
2024-04-03 12:26:51 -07:00 |
|
Krrish Dholakia
|
1d341970ba
|
feat(vertex_ai_anthropic.py): add claude 3 on vertex ai support - working .completions call
.completions() call works
|
2024-04-02 22:07:39 -07:00 |
|
Ishaan Jaff
|
4d76ec43ac
|
Merge pull request #2808 from BerriAI/litellm_use_all_proxy_team_models_auth
[feat] use `all-proxy-models` and `all-team-models` with Admin UI
|
2024-04-02 21:48:30 -07:00 |
|
Krrish Dholakia
|
b5ca4cc235
|
test(test_update_spend.py): fix test with right init
|
2024-04-02 21:11:26 -07:00 |
|
Ishaan Jaff
|
afd81f1609
|
test new team request
|
2024-04-02 20:52:16 -07:00 |
|
CLARKBENHAM
|
44cb0f352a
|
formating
|
2024-04-02 19:56:07 -07:00 |
|
CLARKBENHAM
|
164898a213
|
fix lowest latency tests
|
2024-04-02 19:10:40 -07:00 |
|
CLARKBENHAM
|
29573b0967
|
param both tests to include failure (also fix prev)
|
2024-04-02 18:53:42 -07:00 |
|
Krrish Dholakia
|
d7601a4844
|
perf(proxy_server.py): batch write spend logs
reduces prisma client errors, by batch writing spend logs - max 1k logs at a time
|
2024-04-02 18:46:55 -07:00 |
|
CLARKBENHAM
|
4f95966475
|
tests showing error
|
2024-04-02 18:45:05 -07:00 |
|
Ishaan Jaff
|
21379eb56d
|
Merge pull request #2801 from BerriAI/litellm_support_all_models_as_a_ui_alias
[UI] use all_models alias
|
2024-04-02 17:53:25 -07:00 |
|