Commit graph

1273 commits

Author SHA1 Message Date
Ishaan Jaff
bf909a89f8
Merge pull request #3585 from BerriAI/litellm_router_batch_comp
[Litellm Proxy + litellm.Router] - Pass the same message/prompt to N models
2024-05-11 13:51:45 -07:00
Ishaan Jaff
b8c7bbcb9f support batch /chat/completions on proxy 2024-05-11 13:24:25 -07:00
Krish Dholakia
bbe1300c5b
Merge branch 'main' into feat/add-azure-content-filter 2024-05-11 09:30:38 -07:00
Ishaan Jaff
933f8ed16b fix - proxy_server.py 2024-05-10 13:47:35 -07:00
Nick Wong
d3a228d03b
added changes from upstream
Merge branch 'main' into fix/error-on-get-user-role
2024-05-09 16:14:14 -07:00
Nick Wong
c42f1ce2c6
removed extra default dict return, which causes error if user_role is a string 2024-05-09 16:13:26 -07:00
Krrish Dholakia
927d36148f feat(proxy_server.py): expose new /team/list endpoint
Closes https://github.com/BerriAI/litellm/issues/3523
2024-05-09 13:21:00 -07:00
Krrish Dholakia
e3f25a4a1f fix(auth_checks.py): fix 'get_end_user_object'
await cache get
2024-05-09 13:05:56 -07:00
Krish Dholakia
8ad979cdfe
Merge branch 'main' into litellm_region_based_routing 2024-05-08 22:19:51 -07:00
Krrish Dholakia
3d18897d69 feat(router.py): enable filtering model group by 'allowed_model_region' 2024-05-08 22:10:17 -07:00
Krrish Dholakia
db666b01e5 feat(proxy_server.py): add CRUD endpoints for 'end_user' management
allow admin to specify region + default models for end users
2024-05-08 18:50:36 -07:00
Ishaan Jaff
6d955ef457
Merge branch 'main' into ui_show_spend_end_user 2024-05-08 18:29:25 -07:00
Ishaan Jaff
a38d9e35fd feat - get price by end_user 2024-05-08 18:19:27 -07:00
Ishaan Jaff
1eea4d1c90 fix /global/spend/end_users 2024-05-08 17:03:38 -07:00
Krrish Dholakia
6575143460 feat(proxy_server.py): return litellm version in response headers 2024-05-08 16:00:08 -07:00
Krish Dholakia
91bb7cd261
Merge pull request #3437 from msabramo/add-engines-model-chat-completions-endpoint
Add `/engines/{model}/chat/completions` endpoint
2024-05-08 14:30:39 -07:00
Krish Dholakia
0e709fdc21
Merge branch 'main' into litellm_ui_fixes_6 2024-05-07 22:01:04 -07:00
Krrish Dholakia
fbcda918de feat(ui/model_dashboard.tsx): show if model is config or db model 2024-05-07 21:34:18 -07:00
Krrish Dholakia
5a16bec6a1 feat(model_dashboard.tsx): allow user to edit input cost per token for model on ui
also contains fixes for `/model/update`
2024-05-07 20:57:21 -07:00
Krrish Dholakia
312249ca44 feat(ui/model_dashboard.tsx): show if model is config or db model 2024-05-07 18:29:14 -07:00
Krish Dholakia
2aaaa5e1b4
Merge pull request #3506 from BerriAI/litellm_reintegrate_langfuse_url_slack_alert
feat(slack_alerting.py): reintegrate langfuse trace url for slack alerts
2024-05-07 15:03:29 -07:00
Krrish Dholakia
f210318bf1 fix(proxy_server.py): return budget duration in user response object 2024-05-07 13:47:32 -07:00
Krrish Dholakia
f2766fddbf fix(proxy_server.py): fix /v1/models bug where it would return empty list
handle 'all-team-models' being set for a given key
2024-05-07 13:43:15 -07:00
Krrish Dholakia
872470ff1f feat(slack_alerting.py): reintegrate langfuse trace url for slack alerts
this ensures langfuse trace url returned in llm api exception err
2024-05-07 12:58:49 -07:00
Ishaan Jaff
bfef424b39 fix don't let slack alert block /model/new 2024-05-06 20:47:29 -07:00
Ishaan Jaff
eb84c69ec6 fix - /model/new 2024-05-06 20:45:17 -07:00
Krish Dholakia
aa62d891a0
Merge branch 'main' into litellm_slack_daily_reports 2024-05-06 19:31:20 -07:00
Krrish Dholakia
26c0ed0f2d refactor(proxy_server.py): show ttl's on a top-level enum
Addresses - https://github.com/BerriAI/litellm/issues/2649#issuecomment-2097203372
2024-05-06 18:43:42 -07:00
Krrish Dholakia
6b9b4f05ba feat(proxy_server.py): schedule slack daily report if enabled
if user enabled daily_reports, send them a slack report every 12 hours
2024-05-06 18:25:48 -07:00
Ishaan Jaff
c600371e6e feat - send alert on adding new model 2024-05-06 15:45:07 -07:00
Ishaan Jaff
562ef2d2e1 fix - add better debugging on num_callbacks test 2024-05-06 13:42:20 -07:00
Ishaan Jaff
fccdb92c6b fix - select startTime and endTime on UI 2024-05-03 21:20:19 -07:00
Marc Abramowitz
eb433bde86 Add route: "/engines/{model:path}/chat/completions"
Without this, it results in:

```pytb
Traceback (most recent call last):
  File "/Users/abramowi/Code/OpenSource/litellm/litellm/proxy/proxy_server.py", line 3836, in completion
    raise HTTPException(
fastapi.exceptions.HTTPException: 400: {'error': 'completion: Invalid model name passed in model=gpt-3.5-turbo/chat'}
```
2024-05-03 18:02:29 -07:00
Ishaan Jaff
e7034ea53d feat - filter exceptions by model group 2024-05-03 16:54:24 -07:00
Ishaan Jaff
3dd1e8dfe7
Merge pull request #3427 from BerriAI/litellm_test_alert_size
[Test] - Ensure only 1 slack callback + Size of of all callbacks do not grow
2024-05-03 16:27:16 -07:00
Krish Dholakia
1b35a75245
Merge pull request #3430 from BerriAI/litellm_return_api_base
feat(proxy_server.py): return api base in response headers
2024-05-03 16:25:21 -07:00
Krrish Dholakia
5b39f8e282 feat(proxy_server.py): return api base in response headers
Closes https://github.com/BerriAI/litellm/issues/2631
2024-05-03 15:27:32 -07:00
Ishaan Jaff
ab27866b6a fix test slack alerting len 2024-05-03 14:58:11 -07:00
Ishaan Jaff
3997ea6442 fix - return num callbacks in /active/callbacks 2024-05-03 14:24:01 -07:00
Ishaan Jaff
e99edaf4e1
Merge pull request #3426 from BerriAI/litellm_set_db_exceptions_on_ui
UI - set DB Exceptions webhook_url on UI
2024-05-03 14:05:37 -07:00
Ishaan Jaff
776f541f6c fix bug where slack would get inserting several times 2024-05-03 14:04:38 -07:00
Ishaan Jaff
23d334fe60 proxy - return num callbacks on /health/readiness 2024-05-03 09:14:32 -07:00
Marc Abramowitz
988c37fda3 Disambiguate invalid model name errors
because that error can be thrown in several different places, so
knowing the function it's being thrown from can be very useul for debugging.
2024-05-02 15:02:54 -07:00
Lunik
6cec252b07
feat: Add Azure Content-Safety Proxy hooks
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-02 23:21:08 +02:00
Krish Dholakia
762a1fbd50
Merge pull request #3375 from msabramo/GH-3372
Fix route `/openai/deployments/{model}/chat/completions` not working properly
2024-05-02 13:00:25 -07:00
Krish Dholakia
fffbb73465
Merge branch 'main' into litellm_openmeter_integration 2024-05-01 21:19:29 -07:00
Krrish Dholakia
cdd3e1eef3 build(ui): enable adding openmeter via proxy ui 2024-05-01 21:16:23 -07:00
Ishaan Jaff
26eda88b26 feat - show slow count and total count 2024-05-01 17:18:14 -07:00
Ishaan Jaff
f48f4a767c feat - return slow responses on admin UI 2024-05-01 17:16:33 -07:00
Ishaan Jaff
e9dd4bbe57 fix - dont show cache hits on model latency tracker 2024-05-01 16:51:15 -07:00