Marc Abramowitz
dd166680d1
Move chat_completions before completions
...
so that the `chat_completions` route is defined before the `completions` route.
This is necessary because the `chat_completions` route is more
specific than the `completions` route, and the order of route definitions
matters in FastAPI.
Without this, doing a request to
`/openai/deployments/{model_in_url}/chat/completions` might trigger
`completions` being called (with `model` set to `{model_in_url}/chat` instead of
`chat_completions` getting called, which is the correct function.
Fixes: GH-3372
2024-04-30 15:07:10 -07:00
Ishaan Jaff
1f4f1c6f70
stash /model/metrics/exceptions endpoints
2024-04-30 14:19:23 -07:00
Ishaan Jaff
4b8fda4ac4
log startTime and EndTime for exceptions
2024-04-30 13:34:14 -07:00
Ishaan Jaff
3aad034a8b
feat log request kwargs in error logs
2024-04-30 13:28:26 -07:00
Ishaan Jaff
ad5fddef15
fix log model_group
2024-04-30 13:11:09 -07:00
Ishaan Jaff
ee2a2ce559
fix - log api_base in errors
2024-04-30 13:02:42 -07:00
Ishaan Jaff
06804bc70a
fix - working exception writing
2024-04-30 12:48:17 -07:00
Ishaan Jaff
22725bd44d
fix types for errorLog
2024-04-30 12:31:33 -07:00
Ishaan Jaff
ac1cabe963
add LiteLLM_ErrorLogs to types
2024-04-30 12:16:03 -07:00
Krrish Dholakia
5fe0f38558
docs(load_test.md): load test multiple instances of the proxy w/ tpm/rpm limits on deployments
2024-04-29 15:58:14 -07:00
Krrish Dholakia
7b617e666d
fix(proxy_server.py): return more detailed auth error message.
2024-04-29 07:24:19 -07:00
CyanideByte
82be9a7e67
Merge branch 'BerriAI:main' into main
2024-04-27 20:51:33 -07:00
CyanideByte
03a43b99a5
Added _types.py cases from edwinjosegeorge PR#3340
2024-04-27 20:42:54 -07:00
Ishaan Jaff
de8f928bdd
ui - new build
2024-04-27 17:28:30 -07:00
Krrish Dholakia
d9e0d7ce52
test: replace flaky endpoint
2024-04-27 16:37:09 -07:00
Ishaan Jaff
e49fe47d2e
fix - only run global_proxy_spend on chat completion calls
2024-04-27 14:11:00 -07:00
Krish Dholakia
1a06f009d1
Merge branch 'main' into litellm_default_router_retries
2024-04-27 11:21:57 -07:00
Krrish Dholakia
e05764bdb7
fix(router.py): add /v1/
if missing to base url, for openai-compatible api's
...
Fixes https://github.com/BerriAI/litellm/issues/2279
2024-04-26 17:05:07 -07:00
Krish Dholakia
4b0f73500f
Merge branch 'main' into litellm_default_router_retries
2024-04-26 14:52:24 -07:00
Krrish Dholakia
5583197d63
fix(proxy_server.py): fix setting offset-aware datetime
2024-04-25 21:18:32 -07:00
Ishaan Jaff
1bb82ef42f
ui -new build
2024-04-25 20:33:02 -07:00
Krish Dholakia
40b6b4794b
Merge pull request #3310 from BerriAI/litellm_langfuse_error_logging_2
...
fix(proxy/utils.py): log rejected proxy requests to langfuse
2024-04-25 19:49:59 -07:00
Krrish Dholakia
885de2e3c6
fix(proxy/utils.py): log rejected proxy requests to langfuse
2024-04-25 19:26:27 -07:00
Krish Dholakia
69280177a3
Merge pull request #3308 from BerriAI/litellm_fix_streaming_n
...
fix(utils.py): fix the response object returned when n>1 for stream=true
2024-04-25 18:36:54 -07:00
Ishaan Jaff
72cca34739
ui - new build
2024-04-25 16:39:05 -07:00
Ishaan Jaff
96921864dc
fixes for testing alerting
2024-04-25 16:33:55 -07:00
Ishaan Jaff
6d43d65e22
fix add alert_to_webhook_url to general settings
2024-04-25 15:43:54 -07:00
Ishaan Jaff
61f48aba6f
backend - update slack alert_to_webhook_url_map
2024-04-25 13:47:52 -07:00
Krrish Dholakia
6c5c7cca3d
fix(utils.py): fix the response object returned when n>1 for stream=true
...
Fixes https://github.com/BerriAI/litellm/issues/3276
2024-04-25 13:27:29 -07:00
Ishaan Jaff
1d5e70f7a0
pass alert type on alerting handle
2024-04-25 13:05:34 -07:00
Krrish Dholakia
160acc085a
fix(router.py): fix default retry logic
2024-04-25 11:57:27 -07:00
Krrish Dholakia
4f46b4c397
fix(factory.py): add replicate meta llama prompt templating support
2024-04-25 08:25:00 -07:00
Krrish Dholakia
b8f862bb76
fix(proxy_server.py): fix update router
2024-04-24 23:01:21 -07:00
Krrish Dholakia
fe188f3cc1
fix(proxy_server.py): fix updating non-router settings for proxy config
2024-04-24 22:50:04 -07:00
Krish Dholakia
435a4b5ed4
Merge pull request #3267 from BerriAI/litellm_openai_streaming_fix
...
fix(utils.py): fix streaming to not return usage dict
2024-04-24 21:08:33 -07:00
Ishaan Jaff
70ca65d1ef
ui - new build
2024-04-24 21:04:44 -07:00
Ishaan Jaff
b84502c2aa
ui - new build
2024-04-24 20:40:01 -07:00
Krrish Dholakia
495aebb582
fix(utils.py): fix setattr error
2024-04-24 20:19:27 -07:00
Krrish Dholakia
5650e8ea44
feat(router.py): support mock testing fallbacks flag
...
allow user to test if fallbacks work as expected with a `mock_testing_fallbacks = True` flag set during a call
2024-04-24 20:13:10 -07:00
Krrish Dholakia
f54510b6ee
fix(proxy_server.py): fix /config/update
/
...
allows updating router config via UI and having the change be propogated across all proxy instances by persisting config changes to the db
2024-04-24 16:42:42 -07:00
Ishaan Jaff
2ac3885a50
Merge pull request #3277 from BerriAI/litellm_update_deployments
...
[UI] V0 - Edit Model tpm, rpm, api_base
2024-04-24 14:03:00 -07:00
Krrish Dholakia
1f7abb8d92
fix(langfuse.py): fix langfuse environ check
2024-04-24 13:21:00 -07:00
Krrish Dholakia
57cd3f482a
bump: version 1.35.23 → 1.35.24
2024-04-24 11:13:37 -07:00
Krrish Dholakia
3fdb2f0779
fix(ui): fix add slack/langfuse logging on ui
2024-04-24 10:58:44 -07:00
Ishaan Jaff
efbf85a5ad
/model/update endpoint
2024-04-24 10:39:20 -07:00
Krrish Dholakia
26e9ae38ce
fix(proxy_server.py): add new flag for disable sharing master key on ui
2024-04-24 10:06:01 -07:00
Krrish Dholakia
f800ebc3c1
bump: version 1.35.22 → 1.35.23
2024-04-24 08:16:33 -07:00
Krrish Dholakia
a77bb2da5d
build(ui): add ui updates to main build
2024-04-23 22:53:37 -07:00
Ishaan Jaff
aa08397e34
ui - new build
2024-04-23 22:23:49 -07:00
Ishaan Jaff
1a785f255e
Merge pull request #3259 from BerriAI/litellm_filter_metrics_model_group
...
UI - filter metrics by model group
2024-04-23 22:22:05 -07:00