Ishaan Jaff
|
adf3e90f45
|
ui - new build
|
2024-05-01 13:32:32 -07:00 |
|
Ishaan Jaff
|
b3a788142b
|
Merge pull request #3380 from BerriAI/ui_polish_viewing_model_latencies
[UI] Polish viewing Model Latencies
|
2024-05-01 09:44:53 -07:00 |
|
Ishaan Jaff
|
94b98f5c4e
|
clean up model latency metrics
|
2024-05-01 08:27:01 -07:00 |
|
Krrish Dholakia
|
d0f9f8c0ed
|
fix(proxy/utils.py): emit number of spend transactions for keys being written to db in a batch
|
2024-05-01 08:25:04 -07:00 |
|
Ishaan Jaff
|
fc5a845838
|
fix - prisma schema
|
2024-04-30 23:09:53 -07:00 |
|
Ishaan Jaff
|
1e94d53a9b
|
(ui - new build)
|
2024-04-30 22:54:51 -07:00 |
|
Ishaan Jaff
|
b9238a00af
|
ui - show tokens / sec
|
2024-04-30 22:44:28 -07:00 |
|
Ishaan Jaff
|
0c464f7f61
|
fix - viewing model metrics
|
2024-04-30 18:26:14 -07:00 |
|
Ishaan Jaff
|
f2849d0641
|
fix - track litellm_model_name in LiteLLM_ErrorLogs
|
2024-04-30 17:31:40 -07:00 |
|
Ishaan Jaff
|
8a1a043801
|
backend - show model latency per token
|
2024-04-30 17:23:36 -07:00 |
|
Ishaan Jaff
|
a2a8fef8f4
|
fix passing starttime and endtime to model/exceptions
|
2024-04-30 16:53:53 -07:00 |
|
Ishaan Jaff
|
26a5d85869
|
fix - backend return exceptions
|
2024-04-30 15:41:16 -07:00 |
|
Ishaan Jaff
|
1f4f1c6f70
|
stash /model/metrics/exceptions endpoints
|
2024-04-30 14:19:23 -07:00 |
|
Ishaan Jaff
|
4b8fda4ac4
|
log startTime and EndTime for exceptions
|
2024-04-30 13:34:14 -07:00 |
|
Ishaan Jaff
|
3aad034a8b
|
feat log request kwargs in error logs
|
2024-04-30 13:28:26 -07:00 |
|
Ishaan Jaff
|
ad5fddef15
|
fix log model_group
|
2024-04-30 13:11:09 -07:00 |
|
Ishaan Jaff
|
ee2a2ce559
|
fix - log api_base in errors
|
2024-04-30 13:02:42 -07:00 |
|
Ishaan Jaff
|
06804bc70a
|
fix - working exception writing
|
2024-04-30 12:48:17 -07:00 |
|
Ishaan Jaff
|
22725bd44d
|
fix types for errorLog
|
2024-04-30 12:31:33 -07:00 |
|
Ishaan Jaff
|
ac1cabe963
|
add LiteLLM_ErrorLogs to types
|
2024-04-30 12:16:03 -07:00 |
|
Krrish Dholakia
|
5fe0f38558
|
docs(load_test.md): load test multiple instances of the proxy w/ tpm/rpm limits on deployments
|
2024-04-29 15:58:14 -07:00 |
|
Krrish Dholakia
|
7b617e666d
|
fix(proxy_server.py): return more detailed auth error message.
|
2024-04-29 07:24:19 -07:00 |
|
CyanideByte
|
82be9a7e67
|
Merge branch 'BerriAI:main' into main
|
2024-04-27 20:51:33 -07:00 |
|
CyanideByte
|
03a43b99a5
|
Added _types.py cases from edwinjosegeorge PR#3340
|
2024-04-27 20:42:54 -07:00 |
|
Ishaan Jaff
|
de8f928bdd
|
ui - new build
|
2024-04-27 17:28:30 -07:00 |
|
Krrish Dholakia
|
d9e0d7ce52
|
test: replace flaky endpoint
|
2024-04-27 16:37:09 -07:00 |
|
Ishaan Jaff
|
e49fe47d2e
|
fix - only run global_proxy_spend on chat completion calls
|
2024-04-27 14:11:00 -07:00 |
|
Krish Dholakia
|
1a06f009d1
|
Merge branch 'main' into litellm_default_router_retries
|
2024-04-27 11:21:57 -07:00 |
|
Krrish Dholakia
|
e05764bdb7
|
fix(router.py): add /v1/ if missing to base url, for openai-compatible api's
Fixes https://github.com/BerriAI/litellm/issues/2279
|
2024-04-26 17:05:07 -07:00 |
|
Krish Dholakia
|
4b0f73500f
|
Merge branch 'main' into litellm_default_router_retries
|
2024-04-26 14:52:24 -07:00 |
|
Krrish Dholakia
|
5583197d63
|
fix(proxy_server.py): fix setting offset-aware datetime
|
2024-04-25 21:18:32 -07:00 |
|
Ishaan Jaff
|
1bb82ef42f
|
ui -new build
|
2024-04-25 20:33:02 -07:00 |
|
Krish Dholakia
|
40b6b4794b
|
Merge pull request #3310 from BerriAI/litellm_langfuse_error_logging_2
fix(proxy/utils.py): log rejected proxy requests to langfuse
|
2024-04-25 19:49:59 -07:00 |
|
Krrish Dholakia
|
885de2e3c6
|
fix(proxy/utils.py): log rejected proxy requests to langfuse
|
2024-04-25 19:26:27 -07:00 |
|
Krish Dholakia
|
69280177a3
|
Merge pull request #3308 from BerriAI/litellm_fix_streaming_n
fix(utils.py): fix the response object returned when n>1 for stream=true
|
2024-04-25 18:36:54 -07:00 |
|
Ishaan Jaff
|
72cca34739
|
ui - new build
|
2024-04-25 16:39:05 -07:00 |
|
Ishaan Jaff
|
96921864dc
|
fixes for testing alerting
|
2024-04-25 16:33:55 -07:00 |
|
Ishaan Jaff
|
6d43d65e22
|
fix add alert_to_webhook_url to general settings
|
2024-04-25 15:43:54 -07:00 |
|
Ishaan Jaff
|
61f48aba6f
|
backend - update slack alert_to_webhook_url_map
|
2024-04-25 13:47:52 -07:00 |
|
Krrish Dholakia
|
6c5c7cca3d
|
fix(utils.py): fix the response object returned when n>1 for stream=true
Fixes https://github.com/BerriAI/litellm/issues/3276
|
2024-04-25 13:27:29 -07:00 |
|
Ishaan Jaff
|
1d5e70f7a0
|
pass alert type on alerting handle
|
2024-04-25 13:05:34 -07:00 |
|
Krrish Dholakia
|
160acc085a
|
fix(router.py): fix default retry logic
|
2024-04-25 11:57:27 -07:00 |
|
Krrish Dholakia
|
4f46b4c397
|
fix(factory.py): add replicate meta llama prompt templating support
|
2024-04-25 08:25:00 -07:00 |
|
Krrish Dholakia
|
b8f862bb76
|
fix(proxy_server.py): fix update router
|
2024-04-24 23:01:21 -07:00 |
|
Krrish Dholakia
|
fe188f3cc1
|
fix(proxy_server.py): fix updating non-router settings for proxy config
|
2024-04-24 22:50:04 -07:00 |
|
Krish Dholakia
|
435a4b5ed4
|
Merge pull request #3267 from BerriAI/litellm_openai_streaming_fix
fix(utils.py): fix streaming to not return usage dict
|
2024-04-24 21:08:33 -07:00 |
|
Ishaan Jaff
|
70ca65d1ef
|
ui - new build
|
2024-04-24 21:04:44 -07:00 |
|
Ishaan Jaff
|
b84502c2aa
|
ui - new build
|
2024-04-24 20:40:01 -07:00 |
|
Krrish Dholakia
|
495aebb582
|
fix(utils.py): fix setattr error
|
2024-04-24 20:19:27 -07:00 |
|
Krrish Dholakia
|
5650e8ea44
|
feat(router.py): support mock testing fallbacks flag
allow user to test if fallbacks work as expected with a `mock_testing_fallbacks = True` flag set during a call
|
2024-04-24 20:13:10 -07:00 |
|