Ishaan Jaff
b71f35de72
Merge pull request #3585 from BerriAI/litellm_router_batch_comp
...
[Litellm Proxy + litellm.Router] - Pass the same message/prompt to N models
2024-05-11 13:51:45 -07:00
Ishaan Jaff
c49033a687
support batch /chat/completions on proxy
2024-05-11 13:24:25 -07:00
Krish Dholakia
30117317a4
Merge branch 'main' into feat/add-azure-content-filter
2024-05-11 09:30:38 -07:00
Ishaan Jaff
3b2e295337
fix - proxy_server.py
2024-05-10 13:47:35 -07:00
Nick Wong
a1821428d1
added changes from upstream
...
Merge branch 'main' into fix/error-on-get-user-role
2024-05-09 16:14:14 -07:00
Nick Wong
ea0856408b
removed extra default dict return, which causes error if user_role is a string
2024-05-09 16:13:26 -07:00
Krrish Dholakia
8bb710b9c1
feat(proxy_server.py): expose new /team/list
endpoint
...
Closes https://github.com/BerriAI/litellm/issues/3523
2024-05-09 13:21:00 -07:00
Krrish Dholakia
29ef2261f9
fix(auth_checks.py): fix 'get_end_user_object'
...
await cache get
2024-05-09 13:05:56 -07:00
Krish Dholakia
64ca2fde53
Merge branch 'main' into litellm_region_based_routing
2024-05-08 22:19:51 -07:00
Krrish Dholakia
0ea8222508
feat(router.py): enable filtering model group by 'allowed_model_region'
2024-05-08 22:10:17 -07:00
Krrish Dholakia
7c99808529
feat(proxy_server.py): add CRUD endpoints for 'end_user' management
...
allow admin to specify region + default models for end users
2024-05-08 18:50:36 -07:00
Ishaan Jaff
eeece6e90b
Merge branch 'main' into ui_show_spend_end_user
2024-05-08 18:29:25 -07:00
Ishaan Jaff
41fd17dc3d
feat - get price by end_user
2024-05-08 18:19:27 -07:00
Ishaan Jaff
bfa693e829
fix /global/spend/end_users
2024-05-08 17:03:38 -07:00
Krrish Dholakia
5f93cae3ff
feat(proxy_server.py): return litellm version in response headers
2024-05-08 16:00:08 -07:00
Krish Dholakia
269d6cb16f
Merge pull request #3437 from msabramo/add-engines-model-chat-completions-endpoint
...
Add `/engines/{model}/chat/completions` endpoint
2024-05-08 14:30:39 -07:00
Krish Dholakia
1eb75273cf
Merge branch 'main' into litellm_ui_fixes_6
2024-05-07 22:01:04 -07:00
Krrish Dholakia
ae442f895b
feat(ui/model_dashboard.tsx): show if model is config or db model
2024-05-07 21:34:18 -07:00
Krrish Dholakia
b9ec7acb08
feat(model_dashboard.tsx): allow user to edit input cost per token for model on ui
...
also contains fixes for `/model/update`
2024-05-07 20:57:21 -07:00
Krrish Dholakia
1882ee1c4c
feat(ui/model_dashboard.tsx): show if model is config or db model
2024-05-07 18:29:14 -07:00
Krish Dholakia
8b87500620
Merge pull request #3506 from BerriAI/litellm_reintegrate_langfuse_url_slack_alert
...
feat(slack_alerting.py): reintegrate langfuse trace url for slack alerts
2024-05-07 15:03:29 -07:00
Krrish Dholakia
17a7d4f602
fix(proxy_server.py): return budget duration in user response object
2024-05-07 13:47:32 -07:00
Krrish Dholakia
f503929918
fix(proxy_server.py): fix /v1/models
bug where it would return empty list
...
handle 'all-team-models' being set for a given key
2024-05-07 13:43:15 -07:00
Krrish Dholakia
e5c57cdddb
feat(slack_alerting.py): reintegrate langfuse trace url for slack alerts
...
this ensures langfuse trace url returned in llm api exception err
2024-05-07 12:58:49 -07:00
Ishaan Jaff
8608795fc5
fix don't let slack alert block /model/new
2024-05-06 20:47:29 -07:00
Ishaan Jaff
43cd3f05d2
fix - /model/new
2024-05-06 20:45:17 -07:00
Krish Dholakia
185581d5a9
Merge branch 'main' into litellm_slack_daily_reports
2024-05-06 19:31:20 -07:00
Krrish Dholakia
3cada96f22
refactor(proxy_server.py): show ttl's on a top-level enum
...
Addresses - https://github.com/BerriAI/litellm/issues/2649#issuecomment-2097203372
2024-05-06 18:43:42 -07:00
Krrish Dholakia
2170e1723d
feat(proxy_server.py): schedule slack daily report if enabled
...
if user enabled daily_reports, send them a slack report every 12 hours
2024-05-06 18:25:48 -07:00
Ishaan Jaff
881313f335
feat - send alert on adding new model
2024-05-06 15:45:07 -07:00
Ishaan Jaff
e5488a94dd
fix - add better debugging on num_callbacks test
2024-05-06 13:42:20 -07:00
Ishaan Jaff
5ffe9a8008
fix - select startTime and endTime on UI
2024-05-03 21:20:19 -07:00
Marc Abramowitz
15ba2977ef
Add route: "/engines/{model:path}/chat/completions"
...
Without this, it results in:
```pytb
Traceback (most recent call last):
File "/Users/abramowi/Code/OpenSource/litellm/litellm/proxy/proxy_server.py", line 3836, in completion
raise HTTPException(
fastapi.exceptions.HTTPException: 400: {'error': 'completion: Invalid model name passed in model=gpt-3.5-turbo/chat'}
```
2024-05-03 18:02:29 -07:00
Ishaan Jaff
ca054f95de
feat - filter exceptions by model group
2024-05-03 16:54:24 -07:00
Ishaan Jaff
a0136af9ec
Merge pull request #3427 from BerriAI/litellm_test_alert_size
...
[Test] - Ensure only 1 slack callback + Size of of all callbacks do not grow
2024-05-03 16:27:16 -07:00
Krish Dholakia
53993c612b
Merge pull request #3430 from BerriAI/litellm_return_api_base
...
feat(proxy_server.py): return api base in response headers
2024-05-03 16:25:21 -07:00
Krrish Dholakia
aca36d8b79
feat(proxy_server.py): return api base in response headers
...
Closes https://github.com/BerriAI/litellm/issues/2631
2024-05-03 15:27:32 -07:00
Ishaan Jaff
af9875f358
fix test slack alerting len
2024-05-03 14:58:11 -07:00
Ishaan Jaff
5b97d56e41
fix - return num callbacks in /active/callbacks
2024-05-03 14:24:01 -07:00
Ishaan Jaff
2c62353ce0
Merge pull request #3426 from BerriAI/litellm_set_db_exceptions_on_ui
...
UI - set DB Exceptions webhook_url on UI
2024-05-03 14:05:37 -07:00
Ishaan Jaff
b334b60c41
fix bug where slack would get inserting several times
2024-05-03 14:04:38 -07:00
Ishaan Jaff
051d4daa08
proxy - return num callbacks on /health/readiness
2024-05-03 09:14:32 -07:00
Marc Abramowitz
24ddba804f
Disambiguate invalid model name errors
...
because that error can be thrown in several different places, so
knowing the function it's being thrown from can be very useul for debugging.
2024-05-02 15:02:54 -07:00
Lunik
3ca174bc57
✨ feat: Add Azure Content-Safety Proxy hooks
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-02 23:21:08 +02:00
Krish Dholakia
1b5efc47e1
Merge pull request #3375 from msabramo/GH-3372
...
Fix route `/openai/deployments/{model}/chat/completions` not working properly
2024-05-02 13:00:25 -07:00
Krish Dholakia
dff4649509
Merge branch 'main' into litellm_openmeter_integration
2024-05-01 21:19:29 -07:00
Krrish Dholakia
0b1382aef4
build(ui): enable adding openmeter via proxy ui
2024-05-01 21:16:23 -07:00
Ishaan Jaff
5ce76a5b0d
feat - show slow count and total count
2024-05-01 17:18:14 -07:00
Ishaan Jaff
1721ab64aa
feat - return slow responses on admin UI
2024-05-01 17:16:33 -07:00
Ishaan Jaff
cb5073da62
fix - dont show cache hits on model latency tracker
2024-05-01 16:51:15 -07:00