Ishaan Jaff
c600371e6e
feat - send alert on adding new model
2024-05-06 15:45:07 -07:00
Ishaan Jaff
562ef2d2e1
fix - add better debugging on num_callbacks test
2024-05-06 13:42:20 -07:00
Ishaan Jaff
59dac1bc7a
ui - new build
2024-05-04 10:55:39 -07:00
Lunik
1639a51f24
🔊 fix: Correctly use verbose logging
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-04 11:04:23 +02:00
Lunik
8783fd4895
✨ feat: Use 8 severity levels for azure content safety
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-04 10:45:39 +02:00
Lunik
cb178723ca
📝 doc: Azure content safety Proxy usage
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-04 10:39:43 +02:00
Ishaan Jaff
fccdb92c6b
fix - select startTime and endTime on UI
2024-05-03 21:20:19 -07:00
Ishaan Jaff
01a11ccced
ui - new build
2024-05-03 20:40:58 -07:00
Marc Abramowitz
eb433bde86
Add route: "/engines/{model:path}/chat/completions"
...
Without this, it results in:
```pytb
Traceback (most recent call last):
File "/Users/abramowi/Code/OpenSource/litellm/litellm/proxy/proxy_server.py", line 3836, in completion
raise HTTPException(
fastapi.exceptions.HTTPException: 400: {'error': 'completion: Invalid model name passed in model=gpt-3.5-turbo/chat'}
```
2024-05-03 18:02:29 -07:00
Ishaan Jaff
e7034ea53d
feat - filter exceptions by model group
2024-05-03 16:54:24 -07:00
Ishaan Jaff
3dd1e8dfe7
Merge pull request #3427 from BerriAI/litellm_test_alert_size
...
[Test] - Ensure only 1 slack callback + Size of of all callbacks do not grow
2024-05-03 16:27:16 -07:00
Krish Dholakia
1b35a75245
Merge pull request #3430 from BerriAI/litellm_return_api_base
...
feat(proxy_server.py): return api base in response headers
2024-05-03 16:25:21 -07:00
Krrish Dholakia
5b39f8e282
feat(proxy_server.py): return api base in response headers
...
Closes https://github.com/BerriAI/litellm/issues/2631
2024-05-03 15:27:32 -07:00
Ishaan Jaff
ab27866b6a
fix test slack alerting len
2024-05-03 14:58:11 -07:00
Ishaan Jaff
3997ea6442
fix - return num callbacks in /active/callbacks
2024-05-03 14:24:01 -07:00
Ishaan Jaff
e99edaf4e1
Merge pull request #3426 from BerriAI/litellm_set_db_exceptions_on_ui
...
UI - set DB Exceptions webhook_url on UI
2024-05-03 14:05:37 -07:00
Ishaan Jaff
776f541f6c
fix bug where slack would get inserting several times
2024-05-03 14:04:38 -07:00
Lunik
9ba9b3891f
⚡ ️ perf: Remove test violation on each stream chunk
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-03 20:51:40 +02:00
Lunik
e7405f105c
✅ ci: Add tests
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-03 20:50:37 +02:00
Ishaan Jaff
23d334fe60
proxy - return num callbacks on /health/readiness
2024-05-03 09:14:32 -07:00
Krrish Dholakia
fdc4fdb91a
fix(proxy/utils.py): fix slack alerting to only raise alerts for llm api exceptions
...
don't spam for bad user requests. Closes https://github.com/BerriAI/litellm/issues/3395
2024-05-02 17:18:21 -07:00
Marc Abramowitz
988c37fda3
Disambiguate invalid model name errors
...
because that error can be thrown in several different places, so
knowing the function it's being thrown from can be very useul for debugging.
2024-05-02 15:02:54 -07:00
Krrish Dholakia
acda064be6
fix(proxy/utils.py): fix retry logic for generic data request
2024-05-02 14:50:50 -07:00
Lunik
6cec252b07
✨ feat: Add Azure Content-Safety Proxy hooks
...
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-05-02 23:21:08 +02:00
Krish Dholakia
762a1fbd50
Merge pull request #3375 from msabramo/GH-3372
...
Fix route `/openai/deployments/{model}/chat/completions` not working properly
2024-05-02 13:00:25 -07:00
Krrish Dholakia
0251543e7a
refactor(main.py): trigger new build
2024-05-01 21:59:33 -07:00
Ishaan Jaff
761aa7e5c8
ui - new build
2024-05-01 21:43:00 -07:00
Krish Dholakia
fffbb73465
Merge branch 'main' into litellm_openmeter_integration
2024-05-01 21:19:29 -07:00
Krrish Dholakia
cdd3e1eef3
build(ui): enable adding openmeter via proxy ui
2024-05-01 21:16:23 -07:00
Ishaan Jaff
00969aa682
ui - new build
2024-05-01 19:52:57 -07:00
Krrish Dholakia
2a9651b3ca
feat(openmeter.py): add support for user billing
...
open-meter supports user based billing. Closes https://github.com/BerriAI/litellm/issues/1268
2024-05-01 17:23:48 -07:00
Ishaan Jaff
26eda88b26
feat - show slow count and total count
2024-05-01 17:18:14 -07:00
Ishaan Jaff
f48f4a767c
feat - return slow responses on admin UI
2024-05-01 17:16:33 -07:00
Ishaan Jaff
e9dd4bbe57
fix - dont show cache hits on model latency tracker
2024-05-01 16:51:15 -07:00
Ishaan Jaff
2b467a847a
fix latency tracking tool tip
2024-05-01 16:47:30 -07:00
Ishaan Jaff
adf3e90f45
ui - new build
2024-05-01 13:32:32 -07:00
Ishaan Jaff
b3a788142b
Merge pull request #3380 from BerriAI/ui_polish_viewing_model_latencies
...
[UI] Polish viewing Model Latencies
2024-05-01 09:44:53 -07:00
Ishaan Jaff
94b98f5c4e
clean up model latency metrics
2024-05-01 08:27:01 -07:00
Krrish Dholakia
d0f9f8c0ed
fix(proxy/utils.py): emit number of spend transactions for keys being written to db in a batch
2024-05-01 08:25:04 -07:00
Ishaan Jaff
fc5a845838
fix - prisma schema
2024-04-30 23:09:53 -07:00
Ishaan Jaff
1e94d53a9b
(ui - new build)
2024-04-30 22:54:51 -07:00
Ishaan Jaff
b9238a00af
ui - show tokens / sec
2024-04-30 22:44:28 -07:00
Ishaan Jaff
0c464f7f61
fix - viewing model metrics
2024-04-30 18:26:14 -07:00
Ishaan Jaff
f2849d0641
fix - track litellm_model_name in LiteLLM_ErrorLogs
2024-04-30 17:31:40 -07:00
Ishaan Jaff
8a1a043801
backend - show model latency per token
2024-04-30 17:23:36 -07:00
Ishaan Jaff
a2a8fef8f4
fix passing starttime and endtime to model/exceptions
2024-04-30 16:53:53 -07:00
Ishaan Jaff
26a5d85869
fix - backend return exceptions
2024-04-30 15:41:16 -07:00
Marc Abramowitz
dd166680d1
Move chat_completions before completions
...
so that the `chat_completions` route is defined before the `completions` route.
This is necessary because the `chat_completions` route is more
specific than the `completions` route, and the order of route definitions
matters in FastAPI.
Without this, doing a request to
`/openai/deployments/{model_in_url}/chat/completions` might trigger
`completions` being called (with `model` set to `{model_in_url}/chat` instead of
`chat_completions` getting called, which is the correct function.
Fixes: GH-3372
2024-04-30 15:07:10 -07:00
Ishaan Jaff
1f4f1c6f70
stash /model/metrics/exceptions endpoints
2024-04-30 14:19:23 -07:00
Ishaan Jaff
4b8fda4ac4
log startTime and EndTime for exceptions
2024-04-30 13:34:14 -07:00