Commit graph

8681 commits

Author SHA1 Message Date
Krrish Dholakia
0fb7afe820 feat(proxy_server.py): working /audio/transcription endpoint 2024-03-08 18:20:27 -08:00
Guillermo
bb427b4659
Update deploy.md 2024-03-09 02:30:17 +01:00
Guillermo
37b4dde7fd
Add quickstart deploy with k8s 2024-03-09 02:24:07 +01:00
ishaan-jaff
0a538fe679 (feat) use no-log to disable per request logging 2024-03-08 16:56:20 -08:00
ishaan-jaff
ddd231a8c2 (feat) use no-log as a litellm param 2024-03-08 16:46:38 -08:00
ishaan-jaff
d6dc28f0ed (fix) proxy setting success callbacks 2024-03-08 16:27:53 -08:00
ishaan-jaff
4ff68c8562 (docs) no log requests 2024-03-08 16:26:25 -08:00
ishaan-jaff
986a526790 (feat) disable logging per request 2024-03-08 16:25:54 -08:00
Ishaan Jaff
ad4bfee3ee
Merge pull request #2406 from BerriAI/litellm_locust_load_test
[Feat] LiteLLM - use cpu_count for default num_workers, run locust load test
2024-03-08 15:41:40 -08:00
ishaan-jaff
9ed51e791b (fix) default num workers 2024-03-08 15:24:08 -08:00
ishaan-jaff
2d71f54afb (docs) load test litellm 2024-03-08 15:18:06 -08:00
Krrish Dholakia
cc0294b2f2 fix(proxy_server.py): fix tagging of endpoints 2024-03-08 14:29:31 -08:00
Krrish Dholakia
8c6d5b7f16 feat(proxy_server.py): supports /audio/transcription endpoint on proxy 2024-03-08 14:28:07 -08:00
Krrish Dholakia
fe125a5131 test(test_whisper.py): add testing for load balancing whisper endpoints on router 2024-03-08 14:19:37 -08:00
Krrish Dholakia
9274245a0b test(test_whisper.py): fix getting path for audio file in test 2024-03-08 14:10:19 -08:00
Krrish Dholakia
6fa585d001 test(test_whisper.py): fix test 2024-03-08 14:02:34 -08:00
Krrish Dholakia
ae54b398d2 feat(router.py): add load balancing for async transcription calls 2024-03-08 13:58:15 -08:00
Krrish Dholakia
321769a74d build(model_prices_and_context_window.json): add pricing for whisper endpoints (openai + azure) 2024-03-08 13:50:38 -08:00
Krrish Dholakia
6b1049217e feat(azure.py): add support for calling whisper endpoints on azure 2024-03-08 13:48:38 -08:00
ishaan-jaff
65ccfc35ca (fix) azure extra setting client 2024-03-08 13:44:39 -08:00
ishaan-jaff
2f6e15655a (feat) set api_version for Azure 2024-03-08 13:38:29 -08:00
ishaan-jaff
d67c63b0c3 (fix) use azure api_version 2024-03-08 13:33:58 -08:00
ishaan-jaff
2aafbe390b (feat) read passed api_version 2024-03-08 13:16:12 -08:00
Krish Dholakia
713f5991b8
Merge pull request #2402 from GuillermoBlasco/patch-1
docs: fix yaml typo in proxy/configs.md
2024-03-08 11:56:37 -08:00
Guillermo
a693045460
docs: fix yaml typo in proxy/configs.md
equals was used instead of : as key-value delimiter in yaml
2024-03-08 20:33:47 +01:00
Ishaan Jaff
85b981f602
Merge pull request #2399 from BerriAI/litellm_support_name_for_anthropic
[Feat] Support messages.name for claude-3, perplexity ai API
2024-03-08 10:47:45 -08:00
Krrish Dholakia
696eb54455 feat(main.py): support openai transcription endpoints
enable user to load balance between openai + azure transcription endpoints
2024-03-08 10:25:19 -08:00
ishaan-jaff
96e3696138 (fix) support name on perplexity/ 2024-03-08 09:41:58 -08:00
ishaan-jaff
f70feb1806 (test) name with claude-3 2024-03-08 09:33:54 -08:00
Krrish Dholakia
f8f01e5224 bump: version 1.30.2 → 1.30.3 2024-03-08 08:12:36 -08:00
Krrish Dholakia
2f9a39f30c refactor(main.py): trigger new build 2024-03-08 08:12:22 -08:00
Krrish Dholakia
0e7b30bec9 fix(utils.py): return function name for ollama_chat function calls 2024-03-08 08:01:10 -08:00
Krrish Dholakia
0cf056f493 fix(proxy_server.py): use argon2 for faster api key checking
0.04s latency boost on load test
2024-03-07 21:48:18 -08:00
ishaan-jaff
b4e12fb8fd (docs) litellm cloud formation stack 2024-03-07 21:01:28 -08:00
ishaan-jaff
4a4d36d923 (feat) bump litellm 2024-03-07 20:45:34 -08:00
Ishaan Jaff
302473eef1
Merge pull request #2391 from BerriAI/litellm_cloudform_stack
(feat) LiteLLM AWS CloudFormation Stack Template
2024-03-07 20:44:46 -08:00
Ishaan Jaff
90d7b61ec2
Merge pull request #2393 from BerriAI/litellm_router_debug_get_dep
[Feat] Load Balancing - View Metrics about selected deployments in server logs
2024-03-07 20:44:18 -08:00
ishaan-jaff
86ac020b12 (fix) show latency per deployment on router debug logs 2024-03-07 18:50:45 -08:00
ishaan-jaff
6f0faca85b (feat) print debug info per deployment 2024-03-07 18:33:09 -08:00
ishaan-jaff
323f15aa2f (fix) litellm cloud formation stack 2024-03-07 18:06:59 -08:00
ishaan-jaff
2f960a9651 (feat) auto-scale 2024-03-07 17:51:12 -08:00
ishaan-jaff
702e46b53d (build) stack with db 2024-03-07 17:50:14 -08:00
ishaan-jaff
da4b150398 (feat) init cloud form stack 2024-03-07 16:49:39 -08:00
Krish Dholakia
dec967f647
Update model_prices_and_context_window.json 2024-03-07 16:31:23 -08:00
Ishaan Jaff
25a24da037
Merge pull request #2390 from BerriAI/litellm_better_debug_prisma_logs
[FIX] Proxy better debug prisma logs
2024-03-07 16:19:26 -08:00
ishaan-jaff
8c3ec15856 (fix) improve improve prisma alerting/debug 2024-03-07 15:44:03 -08:00
Ishaan Jaff
fb791ff0c9
Merge pull request #2387 from DanielChico/feature/issue-2146-fix-discord-publish-workflow
feat: add realease details to discord notification message
2024-03-07 15:34:06 -08:00
ishaan-jaff
5217e800c4 (fix) better prisma debug logs 2024-03-07 14:04:02 -08:00
ishaan-jaff
b64e4c3e6f (feat) better debugging when reading prisma 2024-03-07 13:42:54 -08:00
Ishaan Jaff
1f15c79252
Merge pull request #2388 from BerriAI/litellm_docs_show_how_to_set_lb_config
(docs) setting load balancing config
2024-03-07 12:23:40 -08:00