Commit graph

10714 commits

Author SHA1 Message Date
Ishaan Jaff
f2849d0641 fix - track litellm_model_name in LiteLLM_ErrorLogs 2024-04-30 17:31:40 -07:00
Ishaan Jaff
8a1a043801 backend - show model latency per token 2024-04-30 17:23:36 -07:00
Ishaan Jaff
8177ef5ec0 ui - show model latency / token 2024-04-30 17:23:27 -07:00
Ishaan Jaff
ce1817380e feat ui - modelExceptionsCall 2024-04-30 16:56:45 -07:00
Ishaan Jaff
a2a8fef8f4 fix passing starttime and endtime to model/exceptions 2024-04-30 16:53:53 -07:00
Ishaan Jaff
26a5d85869 fix - backend return exceptions 2024-04-30 15:41:16 -07:00
Ishaan Jaff
49f83ce204 ui - show models analytics 2024-04-30 15:16:25 -07:00
Ishaan Jaff
b9a0a13516 ui - show model usage 2024-04-30 14:28:19 -07:00
Ishaan Jaff
1f4f1c6f70 stash /model/metrics/exceptions endpoints 2024-04-30 14:19:23 -07:00
Ishaan Jaff
0b0be700fc
Merge pull request #3371 from BerriAI/litellm_log_errors_db
[Feat] Write LLM Exception to LiteLLM Proxy DB
2024-04-30 13:36:41 -07:00
Ishaan Jaff
4b8fda4ac4 log startTime and EndTime for exceptions 2024-04-30 13:34:14 -07:00
Ishaan Jaff
3aad034a8b feat log request kwargs in error logs 2024-04-30 13:28:26 -07:00
Ishaan Jaff
ad5fddef15 fix log model_group 2024-04-30 13:11:09 -07:00
Ishaan Jaff
ee2a2ce559 fix - log api_base in errors 2024-04-30 13:02:42 -07:00
Ishaan Jaff
06804bc70a fix - working exception writing 2024-04-30 12:48:17 -07:00
Ishaan Jaff
22725bd44d fix types for errorLog 2024-04-30 12:31:33 -07:00
Ishaan Jaff
c7f979e0fe fix schema error logs 2024-04-30 12:31:19 -07:00
Ishaan Jaff
ac1cabe963 add LiteLLM_ErrorLogs to types 2024-04-30 12:16:03 -07:00
Krrish Dholakia
285a3733a9 test(test_image_generation.py): fix test 2024-04-30 12:14:29 -07:00
Ishaan Jaff
d6f7fa7f4e v0 prisma schema 2024-04-30 11:42:17 -07:00
Krrish Dholakia
398d503590 build(model_prices_and_context_window.json): add bedrock llama3 pricing 2024-04-30 11:36:29 -07:00
Krrish Dholakia
00d1440d0d test(test_image_generation.py): change img model for test - bedrock EOL 2024-04-30 08:55:40 -07:00
Krrish Dholakia
d717fa2588 test(test_tpm_rpm_routing_v2.py): fix test - bump number of iteration s 2024-04-30 08:48:55 -07:00
Krrish Dholakia
1cd24d8906 bump: version 1.35.32 → 1.35.33 2024-04-30 07:20:50 -07:00
Krrish Dholakia
020b175ef4 fix(lowest_tpm_rpm_v2.py): skip if item_tpm is None 2024-04-29 21:34:25 -07:00
Ishaan Jaff
81df36b298 docs - slack alerting 2024-04-29 21:33:03 -07:00
Ishaan Jaff
b1e888edad docs example logging to langfuse 2024-04-29 21:26:27 -07:00
Ishaan Jaff
0cad58f5c6 docs logging to langfuse on proxy 2024-04-29 21:26:15 -07:00
Ishaan Jaff
0c99ae9451 docs - fix kub.yaml config on docs 2024-04-29 21:20:29 -07:00
Krrish Dholakia
b46db8b891 feat(utils.py): json logs for raw request sent by litellm
make it easier to view verbose logs in datadog
2024-04-29 19:21:19 -07:00
Krrish Dholakia
f0e48cdd53 fix(router.py): raise better exception when no deployments are available
Fixes https://github.com/BerriAI/litellm/issues/3355
2024-04-29 18:48:04 -07:00
Krrish Dholakia
1e53c06064 test(test_router_caching.py): remove unstable test
test would fail due to timing issues
2024-04-29 18:37:31 -07:00
Krrish Dholakia
e7b4882e97 fix(router.py): fix high-traffic bug for usage-based-routing-v2 2024-04-29 16:48:01 -07:00
Krish Dholakia
09bae3d8ad
Merge pull request #3351 from elisalimli/main
Fix Cohere tool calling
2024-04-29 16:45:48 -07:00
Krish Dholakia
32534b5e91
Merge pull request #3358 from sumanth13131/usage-based-routing-RPM-fix
usage based routing RPM count fix
2024-04-29 16:45:25 -07:00
Krrish Dholakia
bd79e8b516 docs(langfuse_integration.md): add 'existing_trace_id' to langfuse docs 2024-04-29 16:40:38 -07:00
Krrish Dholakia
853b70aba9 fix(langfuse.py): support 'existing_trace_id' param
allow user to call out a trace as pre-existing, this prevents creating a default trace name, and potentially overwriting past traces
2024-04-29 16:39:17 -07:00
Krrish Dholakia
2cf069befb fix(langfuse.py): don't set default trace_name if trace_id given 2024-04-29 16:39:17 -07:00
Ishaan Jaff
d58dd2cbeb
Merge pull request #3360 from BerriAI/litellm_random_pick_lowest_latency
[Fix] Lowest Latency routing - random pick deployments when all latencies=0
2024-04-29 16:31:32 -07:00
Krrish Dholakia
77f155d158 docs(load_test.md): cleanup docs 2024-04-29 16:27:58 -07:00
Krrish Dholakia
af6a21f27c docs(load_test.md): add multi-instance router load test to docs 2024-04-29 16:25:56 -07:00
Ishaan Jaff
4cb4a7f06d fix - lowest latency routing 2024-04-29 16:02:57 -07:00
Krrish Dholakia
8f830bd948 docs(load_test.md): simplify doc 2024-04-29 16:00:02 -07:00
Krrish Dholakia
fcb83781ec docs(load_test.md): formatting 2024-04-29 15:58:41 -07:00
Krrish Dholakia
5fe0f38558 docs(load_test.md): load test multiple instances of the proxy w/ tpm/rpm limits on deployments 2024-04-29 15:58:14 -07:00
Ishaan Jaff
3b0aa05378 fix lowest latency - routing 2024-04-29 15:51:52 -07:00
Ishaan Jaff
5247d7b6a5 test - lowest latency router 2024-04-29 15:51:01 -07:00
Krrish Dholakia
cef2d95bb4 docs(routing.md): add max parallel requests to router docs 2024-04-29 15:37:48 -07:00
Krrish Dholakia
a978f2d881 fix(lowest_tpm_rpm_v2.py): shuffle deployments with same tpm values 2024-04-29 15:23:47 -07:00
Krrish Dholakia
f10a066d36 fix(lowest_tpm_rpm_v2.py): add more detail to 'No deployments available' error message 2024-04-29 15:04:37 -07:00