Commit graph

6137 commits

Author SHA1 Message Date
Krrish Dholakia
5996617f03 feat(main.py): support auto-infering mode if not set 2024-01-25 20:07:31 -08:00
Krrish Dholakia
ac131b3824 fix(utils.py): completion_cost support for image gen models 2024-01-25 18:08:18 -08:00
ishaan-jaff
bdb7c0a0a7 (ci/cd) docker compose up with ui 2024-01-25 17:13:19 -08:00
ishaan-jaff
90d807bdb0 (ci/cd) build ui, litellm on arm64 2024-01-25 17:13:19 -08:00
Ishaan Jaff
2b1177eaa9
Merge pull request #1625 from BerriAI/litellm_improve_admin_ui
[UI] Admin UI improvements
2024-01-25 16:48:02 -08:00
ishaan-jaff
eba38e169a (fix) undo ui changes 2024-01-25 16:44:46 -08:00
ishaan-jaff
a1067521c3 (ui) view app spend 2024-01-25 16:15:05 -08:00
ishaan-jaff
e56721d6c3 (feat) view spend/logs by user_id, view spend/user by user 2024-01-25 16:12:28 -08:00
Krrish Dholakia
878b0be174 bump: version 1.19.3 → 1.19.4 2024-01-25 16:06:12 -08:00
Krrish Dholakia
76468cc351 refactor: trigger new bump 2024-01-25 16:06:01 -08:00
Krish Dholakia
7271f5d9d5
Merge pull request #1623 from BerriAI/litellm_dynamic_vertex_ai_project
fix(main.py): allow vertex ai project and location to be set in completion() call
2024-01-25 16:04:41 -08:00
Krrish Dholakia
fbab810189 fix(main.py): allow vertex ai project and location to be set in completion() call 2024-01-25 15:00:51 -08:00
ishaan-jaff
06506f65f0 bump: version 1.19.2 → 1.19.3 2024-01-25 14:50:51 -08:00
ishaan-jaff
42270258d1 (docs) new gpt-4-0125-preview 2024-01-25 14:48:56 -08:00
ishaan-jaff
e2d9e40886 (test) gpt-4-0125-preview 2024-01-25 14:42:10 -08:00
ishaan-jaff
4404f22f17 (feat) add gpt-4-0125-preview 2024-01-25 14:41:46 -08:00
ishaan-jaff
b3853c3a53 (docs) new OpenAI embedding models 2024-01-25 14:38:03 -08:00
ishaan-jaff
72790f44da (chore) cleanup testing file 2024-01-25 14:36:11 -08:00
ishaan-jaff
f0a11baf19 (feat) add new OpenAI text-embedding-3 2024-01-25 14:34:15 -08:00
ishaan-jaff
6cbde02cab (test) embedding models 2024-01-25 14:30:49 -08:00
Ishaan Jaff
06a5dbfb5e
Merge pull request #1615 from BerriAI/litellm_alerts_budget_tracking
[Feat] Alerts for Proxy Budgets
2024-01-25 12:00:18 -08:00
ishaan-jaff
ca12e70369 (fix) do nothing if alerting is not switched on 2024-01-25 11:58:55 -08:00
ishaan-jaff
e80d32dcdd (fix) alerting debug statements 2024-01-25 11:56:52 -08:00
ishaan-jaff
6dc9be4d43 (docs) config.yaml 2024-01-25 11:41:35 -08:00
ishaan-jaff
6fb3f8f239 (docs) track max_budget on proxy config.yaml 2024-01-25 11:40:56 -08:00
ishaan-jaff
b3f91844cb (fix) better alert message on budgets 2024-01-25 11:40:20 -08:00
ishaan-jaff
450b0a0ad1 (fix) raise correct error when proxy crossed budget 2024-01-25 11:39:57 -08:00
ishaan-jaff
126b87e3fa (fix) raise exception budget_duration is set and max_budget is Not 2024-01-25 11:32:05 -08:00
ishaan-jaff
3ef2afb0e4 (feat) slack alerting budgets 2024-01-25 11:18:06 -08:00
ishaan-jaff
1ab713c76c (feat) alerts proxy budgets 2024-01-25 10:01:32 -08:00
ishaan-jaff
d328a4bad0 v0 basic structure 2024-01-25 09:58:43 -08:00
Krrish Dholakia
39d5407e67 fix(proxy_server.py): don't set tpm/rpm limits unless set
https://github.com/BerriAI/litellm/issues/1594
2024-01-25 09:53:10 -08:00
Krrish Dholakia
fdb28407bc docs(custom_pricing.md): fix tutorial 2024-01-25 09:36:15 -08:00
Krrish Dholakia
01a84837a7 docs(vertex.md): add vertex ai proxy tutorial to docs 2024-01-25 09:11:11 -08:00
Krrish Dholakia
389347250f docs(users.md): add proxy budget to docs 2024-01-25 09:04:23 -08:00
ishaan-jaff
5b151c8566 (ghcr deploy) use platform 2024-01-25 08:43:38 -08:00
Krrish Dholakia
7e06944d80 test(test_dynamodb_logs.py): skip flaky test 2024-01-24 23:18:45 -08:00
Krrish Dholakia
34154dde02 bump: version 1.19.1 → 1.19.2 2024-01-24 22:19:07 -08:00
Krrish Dholakia
81846ffdec fix(proxy/utils.py): handle item not existing during batch updates 2024-01-24 21:49:47 -08:00
Krrish Dholakia
0752048b81 fix(dynamo_db.py): fix update bug 2024-01-24 21:29:56 -08:00
Krrish Dholakia
8e1157fc92 test(test_keys.py): reset proxy spend 2024-01-24 21:08:09 -08:00
Krrish Dholakia
34c4532e7e fix(proxy_server.py): fix handling none value for existing spend object pt.2 2024-01-24 20:39:00 -08:00
Krrish Dholakia
12b2ad9ed1 fix(proxy_server.py): fix handling none value for existing spend object 2024-01-24 20:37:56 -08:00
Krish Dholakia
fe592aa7ec
Merge pull request #1603 from BerriAI/litellm_global_spend_updates
Litellm global spend updates
2024-01-24 20:20:27 -08:00
Krish Dholakia
6501fdb76e
Merge branch 'main' into litellm_global_spend_updates 2024-01-24 20:20:15 -08:00
Krrish Dholakia
30d615f442 build(proxy_server_config.yaml): add proxy budget to default yaml 2024-01-24 20:12:03 -08:00
Krrish Dholakia
b1864c3d11 bump: version 1.19.0 → 1.19.1 2024-01-24 20:10:14 -08:00
Krrish Dholakia
43f139fafd fix(ollama_chat.py): fix default token counting for ollama chat 2024-01-24 20:09:17 -08:00
Krrish Dholakia
05b4d49882 ci(config.yml): add debug logs 2024-01-24 19:54:37 -08:00
ishaan-jaff
3e59a02dfb (test) test /key/gen with max_budget=None 2024-01-24 19:27:53 -08:00