Ishaan Jaff
|
24e918b10f
|
fix clean up prints from slack alerting
|
2024-04-25 19:36:42 -07:00 |
|
Ishaan Jaff
|
bf92a0b31c
|
fix debugging lowest latency router
|
2024-04-25 19:34:28 -07:00 |
|
Ishaan Jaff
|
737af2b458
|
fix better debugging for latency
|
2024-04-25 11:35:08 -07:00 |
|
Ishaan Jaff
|
787735bb5a
|
fix
|
2024-04-25 11:25:03 -07:00 |
|
Ishaan Jaff
|
984259d420
|
temp - show better debug logs for lowest latency
|
2024-04-25 11:22:52 -07:00 |
|
Krrish Dholakia
|
4f46b4c397
|
fix(factory.py): add replicate meta llama prompt templating support
|
2024-04-25 08:25:00 -07:00 |
|
Ishaan Jaff
|
92f21cba30
|
fix - increase default penalty for lowest latency
|
2024-04-25 07:54:25 -07:00 |
|
Ishaan Jaff
|
74817c560e
|
(ci/cd) run again
|
2024-04-24 23:23:14 -07:00 |
|
Ishaan Jaff
|
4e707af592
|
Revert "fix(router.py): fix max retries on set_client"
This reverts commit 821844c1a3 .
|
2024-04-24 23:19:14 -07:00 |
|
Ishaan Jaff
|
13e0ac64ef
|
(fix) updating router settings
|
2024-04-24 23:09:25 -07:00 |
|
Krrish Dholakia
|
72dffdba39
|
fix(router.py): fix default retry on router
|
2024-04-24 23:06:53 -07:00 |
|
Krrish Dholakia
|
b8f862bb76
|
fix(proxy_server.py): fix update router
|
2024-04-24 23:01:21 -07:00 |
|
Krrish Dholakia
|
fe188f3cc1
|
fix(proxy_server.py): fix updating non-router settings for proxy config
|
2024-04-24 22:50:04 -07:00 |
|
Krrish Dholakia
|
0a9cdf6f9b
|
refactor(main.py): trigger new build
|
2024-04-24 22:04:24 -07:00 |
|
Krrish Dholakia
|
821844c1a3
|
fix(router.py): fix max retries on set_client
|
2024-04-24 22:03:01 -07:00 |
|
Ishaan Jaff
|
242830108c
|
(ci/cd) run again
|
2024-04-24 21:09:49 -07:00 |
|
Krish Dholakia
|
435a4b5ed4
|
Merge pull request #3267 from BerriAI/litellm_openai_streaming_fix
fix(utils.py): fix streaming to not return usage dict
|
2024-04-24 21:08:33 -07:00 |
|
Ishaan Jaff
|
70ca65d1ef
|
ui - new build
|
2024-04-24 21:04:44 -07:00 |
|
Krrish Dholakia
|
dacadbf624
|
fix(utils.py): fix anthropic streaming return usage tokens
|
2024-04-24 20:56:10 -07:00 |
|
Ishaan Jaff
|
2c7f4695d9
|
Merge pull request #3283 from BerriAI/litellm_debug_lowest_latency
[Fix] Add better observability for debugging lowest latency routing
|
2024-04-24 20:42:52 -07:00 |
|
Ishaan Jaff
|
b84502c2aa
|
ui - new build
|
2024-04-24 20:40:01 -07:00 |
|
Krrish Dholakia
|
df7db2b870
|
fix(factory.py): support llama3 instuct chat template
allows automatic templating for llama3 instruct requests
|
2024-04-24 20:35:10 -07:00 |
|
Krrish Dholakia
|
495aebb582
|
fix(utils.py): fix setattr error
|
2024-04-24 20:19:27 -07:00 |
|
Ishaan Jaff
|
ffc277c2bb
|
temp fix _add_langfuse_trace_id_to_alert
|
2024-04-24 20:19:21 -07:00 |
|
Ishaan Jaff
|
212369498e
|
fix - set latency stats in kwargs
|
2024-04-24 20:13:45 -07:00 |
|
Krrish Dholakia
|
5650e8ea44
|
feat(router.py): support mock testing fallbacks flag
allow user to test if fallbacks work as expected with a `mock_testing_fallbacks = True` flag set during a call
|
2024-04-24 20:13:10 -07:00 |
|
Ishaan Jaff
|
24b0dfee29
|
fix - add _get_deployment_latencies_to_alert
|
2024-04-24 20:09:02 -07:00 |
|
Krrish Dholakia
|
94cbe5516e
|
feat(router.py): support mock testing fallbacks flag on router
|
2024-04-24 17:33:00 -07:00 |
|
Krrish Dholakia
|
f54510b6ee
|
fix(proxy_server.py): fix /config/update /
allows updating router config via UI and having the change be propogated across all proxy instances by persisting config changes to the db
|
2024-04-24 16:42:42 -07:00 |
|
Ishaan Jaff
|
2e6fc91a75
|
test - lowest latency logger
|
2024-04-24 16:35:43 -07:00 |
|
Ishaan Jaff
|
bf6abed808
|
feat - penalize timeout errors
|
2024-04-24 16:35:00 -07:00 |
|
Krrish Dholakia
|
bae6f41017
|
build(add-fallbacks-on-UI): allows admin to add fallbacks on the UI
|
2024-04-24 15:40:02 -07:00 |
|
Ishaan Jaff
|
166f3915f9
|
Merge pull request #3279 from BerriAI/litellm_show_api_base_exceptions
[Feat] Show model, api base in APITimeoutError exceptions
|
2024-04-24 14:59:14 -07:00 |
|
Ishaan Jaff
|
7a5c07de22
|
fix - show api_base, model in exceptions
|
2024-04-24 14:03:48 -07:00 |
|
Ishaan Jaff
|
2ac3885a50
|
Merge pull request #3277 from BerriAI/litellm_update_deployments
[UI] V0 - Edit Model tpm, rpm, api_base
|
2024-04-24 14:03:00 -07:00 |
|
Ishaan Jaff
|
ca4fd85296
|
fix show api_base, model in timeout errors
|
2024-04-24 14:01:32 -07:00 |
|
Ishaan Jaff
|
36f8431bf0
|
(ci/cd) testing
|
2024-04-24 13:25:18 -07:00 |
|
Krrish Dholakia
|
1f7abb8d92
|
fix(langfuse.py): fix langfuse environ check
|
2024-04-24 13:21:00 -07:00 |
|
Krish Dholakia
|
263439ee4a
|
Merge pull request #3098 from greenscale-ai/main
Support for Greenscale AI logging
|
2024-04-24 13:09:03 -07:00 |
|
Krrish Dholakia
|
57cd3f482a
|
bump: version 1.35.23 → 1.35.24
|
2024-04-24 11:13:37 -07:00 |
|
Krrish Dholakia
|
3fdb2f0779
|
fix(ui): fix add slack/langfuse logging on ui
|
2024-04-24 10:58:44 -07:00 |
|
Ishaan Jaff
|
efbf85a5ad
|
/model/update endpoint
|
2024-04-24 10:39:20 -07:00 |
|
Krrish Dholakia
|
26e9ae38ce
|
fix(proxy_server.py): add new flag for disable sharing master key on ui
|
2024-04-24 10:06:01 -07:00 |
|
Ishaan Jaff
|
3d1a158b63
|
feat - update deployments
|
2024-04-24 09:53:42 -07:00 |
|
Krrish Dholakia
|
b918f58262
|
fix(vertex_ai.py): raise explicit error when image url fails to download - prevents silent failure
|
2024-04-24 09:23:15 -07:00 |
|
Krrish Dholakia
|
f800ebc3c1
|
bump: version 1.35.22 → 1.35.23
|
2024-04-24 08:16:33 -07:00 |
|
Krrish Dholakia
|
52d5e02d8b
|
fix(prometheus.py): fix user-id get for prometheus
|
2024-04-24 08:08:42 -07:00 |
|
Krrish Dholakia
|
48c2c3d78a
|
fix(utils.py): fix streaming to not return usage dict
Fixes https://github.com/BerriAI/litellm/issues/3237
|
2024-04-24 08:06:07 -07:00 |
|
Krrish Dholakia
|
a77bb2da5d
|
build(ui): add ui updates to main build
|
2024-04-23 22:53:37 -07:00 |
|
Ishaan Jaff
|
aa08397e34
|
ui - new build
|
2024-04-23 22:23:49 -07:00 |
|