Commit graph

1264 commits

Author SHA1 Message Date
Krish Dholakia
64ca2fde53 Merge branch 'main' into litellm_region_based_routing 2024-05-08 22:19:51 -07:00
Krrish Dholakia
0ea8222508 feat(router.py): enable filtering model group by 'allowed_model_region' 2024-05-08 22:10:17 -07:00
Krrish Dholakia
7c99808529 feat(proxy_server.py): add CRUD endpoints for 'end_user' management
allow admin to specify region + default models for end users
2024-05-08 18:50:36 -07:00
Ishaan Jaff
eeece6e90b Merge branch 'main' into ui_show_spend_end_user 2024-05-08 18:29:25 -07:00
Ishaan Jaff
41fd17dc3d feat - get price by end_user 2024-05-08 18:19:27 -07:00
Ishaan Jaff
bfa693e829 fix /global/spend/end_users 2024-05-08 17:03:38 -07:00
Krrish Dholakia
5f93cae3ff feat(proxy_server.py): return litellm version in response headers 2024-05-08 16:00:08 -07:00
Krish Dholakia
269d6cb16f Merge pull request #3437 from msabramo/add-engines-model-chat-completions-endpoint
Add `/engines/{model}/chat/completions` endpoint
2024-05-08 14:30:39 -07:00
Krish Dholakia
1eb75273cf Merge branch 'main' into litellm_ui_fixes_6 2024-05-07 22:01:04 -07:00
Krrish Dholakia
ae442f895b feat(ui/model_dashboard.tsx): show if model is config or db model 2024-05-07 21:34:18 -07:00
Krrish Dholakia
b9ec7acb08 feat(model_dashboard.tsx): allow user to edit input cost per token for model on ui
also contains fixes for `/model/update`
2024-05-07 20:57:21 -07:00
Krrish Dholakia
1882ee1c4c feat(ui/model_dashboard.tsx): show if model is config or db model 2024-05-07 18:29:14 -07:00
Krish Dholakia
8b87500620 Merge pull request #3506 from BerriAI/litellm_reintegrate_langfuse_url_slack_alert
feat(slack_alerting.py): reintegrate langfuse trace url for slack alerts
2024-05-07 15:03:29 -07:00
Krrish Dholakia
17a7d4f602 fix(proxy_server.py): return budget duration in user response object 2024-05-07 13:47:32 -07:00
Krrish Dholakia
f503929918 fix(proxy_server.py): fix /v1/models bug where it would return empty list
handle 'all-team-models' being set for a given key
2024-05-07 13:43:15 -07:00
Krrish Dholakia
e5c57cdddb feat(slack_alerting.py): reintegrate langfuse trace url for slack alerts
this ensures langfuse trace url returned in llm api exception err
2024-05-07 12:58:49 -07:00
Ishaan Jaff
8608795fc5 fix don't let slack alert block /model/new 2024-05-06 20:47:29 -07:00
Ishaan Jaff
43cd3f05d2 fix - /model/new 2024-05-06 20:45:17 -07:00
Krish Dholakia
185581d5a9 Merge branch 'main' into litellm_slack_daily_reports 2024-05-06 19:31:20 -07:00
Krrish Dholakia
3cada96f22 refactor(proxy_server.py): show ttl's on a top-level enum
Addresses - https://github.com/BerriAI/litellm/issues/2649#issuecomment-2097203372
2024-05-06 18:43:42 -07:00
Krrish Dholakia
2170e1723d feat(proxy_server.py): schedule slack daily report if enabled
if user enabled daily_reports, send them a slack report every 12 hours
2024-05-06 18:25:48 -07:00
Ishaan Jaff
881313f335 feat - send alert on adding new model 2024-05-06 15:45:07 -07:00
Ishaan Jaff
e5488a94dd fix - add better debugging on num_callbacks test 2024-05-06 13:42:20 -07:00
Ishaan Jaff
5ffe9a8008 fix - select startTime and endTime on UI 2024-05-03 21:20:19 -07:00
Marc Abramowitz
15ba2977ef Add route: "/engines/{model:path}/chat/completions"
Without this, it results in:

```pytb
Traceback (most recent call last):
  File "/Users/abramowi/Code/OpenSource/litellm/litellm/proxy/proxy_server.py", line 3836, in completion
    raise HTTPException(
fastapi.exceptions.HTTPException: 400: {'error': 'completion: Invalid model name passed in model=gpt-3.5-turbo/chat'}
```
2024-05-03 18:02:29 -07:00
Ishaan Jaff
ca054f95de feat - filter exceptions by model group 2024-05-03 16:54:24 -07:00
Ishaan Jaff
a0136af9ec Merge pull request #3427 from BerriAI/litellm_test_alert_size
[Test] - Ensure only 1 slack callback + Size of of all callbacks do not grow
2024-05-03 16:27:16 -07:00
Krish Dholakia
53993c612b Merge pull request #3430 from BerriAI/litellm_return_api_base
feat(proxy_server.py): return api base in response headers
2024-05-03 16:25:21 -07:00
Krrish Dholakia
aca36d8b79 feat(proxy_server.py): return api base in response headers
Closes https://github.com/BerriAI/litellm/issues/2631
2024-05-03 15:27:32 -07:00
Ishaan Jaff
af9875f358 fix test slack alerting len 2024-05-03 14:58:11 -07:00
Ishaan Jaff
5b97d56e41 fix - return num callbacks in /active/callbacks 2024-05-03 14:24:01 -07:00
Ishaan Jaff
2c62353ce0 Merge pull request #3426 from BerriAI/litellm_set_db_exceptions_on_ui
UI - set DB Exceptions webhook_url on UI
2024-05-03 14:05:37 -07:00
Ishaan Jaff
b334b60c41 fix bug where slack would get inserting several times 2024-05-03 14:04:38 -07:00
Ishaan Jaff
051d4daa08 proxy - return num callbacks on /health/readiness 2024-05-03 09:14:32 -07:00
Marc Abramowitz
24ddba804f Disambiguate invalid model name errors
because that error can be thrown in several different places, so
knowing the function it's being thrown from can be very useul for debugging.
2024-05-02 15:02:54 -07:00
Krish Dholakia
1b5efc47e1 Merge pull request #3375 from msabramo/GH-3372
Fix route `/openai/deployments/{model}/chat/completions` not working properly
2024-05-02 13:00:25 -07:00
Krish Dholakia
dff4649509 Merge branch 'main' into litellm_openmeter_integration 2024-05-01 21:19:29 -07:00
Krrish Dholakia
0b1382aef4 build(ui): enable adding openmeter via proxy ui 2024-05-01 21:16:23 -07:00
Ishaan Jaff
5ce76a5b0d feat - show slow count and total count 2024-05-01 17:18:14 -07:00
Ishaan Jaff
1721ab64aa feat - return slow responses on admin UI 2024-05-01 17:16:33 -07:00
Ishaan Jaff
cb5073da62 fix - dont show cache hits on model latency tracker 2024-05-01 16:51:15 -07:00
Ishaan Jaff
5b70130049 fix latency tracking tool tip 2024-05-01 16:47:30 -07:00
Ishaan Jaff
cf2dd97f0b clean up model latency metrics 2024-05-01 08:27:01 -07:00
Ishaan Jaff
1a2045d63a ui - show tokens / sec 2024-04-30 22:44:28 -07:00
Ishaan Jaff
ee5ebd8f4f fix - viewing model metrics 2024-04-30 18:26:14 -07:00
Ishaan Jaff
69d8f7b89b fix - track litellm_model_name in LiteLLM_ErrorLogs 2024-04-30 17:31:40 -07:00
Ishaan Jaff
5756adc178 backend - show model latency per token 2024-04-30 17:23:36 -07:00
Ishaan Jaff
1c9aa7f0f1 fix passing starttime and endtime to model/exceptions 2024-04-30 16:53:53 -07:00
Ishaan Jaff
672d1c67fc fix - backend return exceptions 2024-04-30 15:41:16 -07:00
Marc Abramowitz
eac4593ab9 Move chat_completions before completions
so that the `chat_completions` route is defined before the `completions` route.
This is necessary because the `chat_completions` route is more
specific than the `completions` route, and the order of route definitions
matters in FastAPI.

Without this, doing a request to
`/openai/deployments/{model_in_url}/chat/completions` might trigger
`completions` being called (with `model` set to `{model_in_url}/chat` instead of
`chat_completions` getting called, which is the correct function.

Fixes: GH-3372
2024-04-30 15:07:10 -07:00