Commit graph

6259 commits

Author SHA1 Message Date
Ishaan Jaff
24e918b10f fix clean up prints from slack alerting 2024-04-25 19:36:42 -07:00
Ishaan Jaff
bf92a0b31c fix debugging lowest latency router 2024-04-25 19:34:28 -07:00
Ishaan Jaff
737af2b458 fix better debugging for latency 2024-04-25 11:35:08 -07:00
Ishaan Jaff
787735bb5a fix 2024-04-25 11:25:03 -07:00
Ishaan Jaff
984259d420 temp - show better debug logs for lowest latency 2024-04-25 11:22:52 -07:00
Krrish Dholakia
4f46b4c397 fix(factory.py): add replicate meta llama prompt templating support 2024-04-25 08:25:00 -07:00
Ishaan Jaff
92f21cba30 fix - increase default penalty for lowest latency 2024-04-25 07:54:25 -07:00
Ishaan Jaff
74817c560e (ci/cd) run again 2024-04-24 23:23:14 -07:00
Ishaan Jaff
4e707af592 Revert "fix(router.py): fix max retries on set_client"
This reverts commit 821844c1a3.
2024-04-24 23:19:14 -07:00
Ishaan Jaff
13e0ac64ef (fix) updating router settings 2024-04-24 23:09:25 -07:00
Krrish Dholakia
72dffdba39 fix(router.py): fix default retry on router 2024-04-24 23:06:53 -07:00
Krrish Dholakia
b8f862bb76 fix(proxy_server.py): fix update router 2024-04-24 23:01:21 -07:00
Krrish Dholakia
fe188f3cc1 fix(proxy_server.py): fix updating non-router settings for proxy config 2024-04-24 22:50:04 -07:00
Krrish Dholakia
0a9cdf6f9b refactor(main.py): trigger new build 2024-04-24 22:04:24 -07:00
Krrish Dholakia
821844c1a3 fix(router.py): fix max retries on set_client 2024-04-24 22:03:01 -07:00
Ishaan Jaff
242830108c (ci/cd) run again 2024-04-24 21:09:49 -07:00
Krish Dholakia
435a4b5ed4
Merge pull request #3267 from BerriAI/litellm_openai_streaming_fix
fix(utils.py): fix streaming to not return usage dict
2024-04-24 21:08:33 -07:00
Ishaan Jaff
70ca65d1ef ui - new build 2024-04-24 21:04:44 -07:00
Krrish Dholakia
dacadbf624 fix(utils.py): fix anthropic streaming return usage tokens 2024-04-24 20:56:10 -07:00
Ishaan Jaff
2c7f4695d9
Merge pull request #3283 from BerriAI/litellm_debug_lowest_latency
[Fix] Add better observability for debugging lowest latency routing
2024-04-24 20:42:52 -07:00
Ishaan Jaff
b84502c2aa ui - new build 2024-04-24 20:40:01 -07:00
Krrish Dholakia
df7db2b870 fix(factory.py): support llama3 instuct chat template
allows automatic templating for llama3 instruct requests
2024-04-24 20:35:10 -07:00
Krrish Dholakia
495aebb582 fix(utils.py): fix setattr error 2024-04-24 20:19:27 -07:00
Ishaan Jaff
ffc277c2bb temp fix _add_langfuse_trace_id_to_alert 2024-04-24 20:19:21 -07:00
Ishaan Jaff
212369498e fix - set latency stats in kwargs 2024-04-24 20:13:45 -07:00
Krrish Dholakia
5650e8ea44 feat(router.py): support mock testing fallbacks flag
allow user to test if fallbacks work as expected with a `mock_testing_fallbacks = True` flag set during a call
2024-04-24 20:13:10 -07:00
Ishaan Jaff
24b0dfee29 fix - add _get_deployment_latencies_to_alert 2024-04-24 20:09:02 -07:00
Krrish Dholakia
94cbe5516e feat(router.py): support mock testing fallbacks flag on router 2024-04-24 17:33:00 -07:00
Krrish Dholakia
f54510b6ee fix(proxy_server.py): fix /config/update/
allows updating router config via UI and having the change be propogated across all proxy instances by persisting config changes to the db
2024-04-24 16:42:42 -07:00
Ishaan Jaff
2e6fc91a75 test - lowest latency logger 2024-04-24 16:35:43 -07:00
Ishaan Jaff
bf6abed808 feat - penalize timeout errors 2024-04-24 16:35:00 -07:00
Krrish Dholakia
bae6f41017 build(add-fallbacks-on-UI): allows admin to add fallbacks on the UI 2024-04-24 15:40:02 -07:00
Ishaan Jaff
166f3915f9
Merge pull request #3279 from BerriAI/litellm_show_api_base_exceptions
[Feat] Show model, api base in APITimeoutError exceptions
2024-04-24 14:59:14 -07:00
Ishaan Jaff
7a5c07de22 fix - show api_base, model in exceptions 2024-04-24 14:03:48 -07:00
Ishaan Jaff
2ac3885a50
Merge pull request #3277 from BerriAI/litellm_update_deployments
[UI] V0 - Edit Model tpm, rpm, api_base
2024-04-24 14:03:00 -07:00
Ishaan Jaff
ca4fd85296 fix show api_base, model in timeout errors 2024-04-24 14:01:32 -07:00
Ishaan Jaff
36f8431bf0 (ci/cd) testing 2024-04-24 13:25:18 -07:00
Krrish Dholakia
1f7abb8d92 fix(langfuse.py): fix langfuse environ check 2024-04-24 13:21:00 -07:00
Krish Dholakia
263439ee4a
Merge pull request #3098 from greenscale-ai/main
Support for Greenscale AI logging
2024-04-24 13:09:03 -07:00
Krrish Dholakia
57cd3f482a bump: version 1.35.23 → 1.35.24 2024-04-24 11:13:37 -07:00
Krrish Dholakia
3fdb2f0779 fix(ui): fix add slack/langfuse logging on ui 2024-04-24 10:58:44 -07:00
Ishaan Jaff
efbf85a5ad /model/update endpoint 2024-04-24 10:39:20 -07:00
Krrish Dholakia
26e9ae38ce fix(proxy_server.py): add new flag for disable sharing master key on ui 2024-04-24 10:06:01 -07:00
Ishaan Jaff
3d1a158b63 feat - update deployments 2024-04-24 09:53:42 -07:00
Krrish Dholakia
b918f58262 fix(vertex_ai.py): raise explicit error when image url fails to download - prevents silent failure 2024-04-24 09:23:15 -07:00
Krrish Dholakia
f800ebc3c1 bump: version 1.35.22 → 1.35.23 2024-04-24 08:16:33 -07:00
Krrish Dholakia
52d5e02d8b fix(prometheus.py): fix user-id get for prometheus 2024-04-24 08:08:42 -07:00
Krrish Dholakia
48c2c3d78a fix(utils.py): fix streaming to not return usage dict
Fixes https://github.com/BerriAI/litellm/issues/3237
2024-04-24 08:06:07 -07:00
Krrish Dholakia
a77bb2da5d build(ui): add ui updates to main build 2024-04-23 22:53:37 -07:00
Ishaan Jaff
aa08397e34 ui - new build 2024-04-23 22:23:49 -07:00