Commit graph

118 commits

Author SHA1 Message Date
Krrish Dholakia
7263ae63c6 fix(proxy_cli.py): allow user to control db connection pool + timeouts from config 2024-02-17 19:03:43 -08:00
ishaan-jaff
8d503caa1e (feat) inc default gunicorn timeout 2024-02-12 21:49:32 -08:00
ishaan-jaff
3b57d285c2 (feat) litellm set ssl_certificates 2024-02-10 17:52:47 -08:00
Krrish Dholakia
4c6bb6a9f6 build(schema.prisma): support direct url on prisma schema 2024-02-08 17:37:37 -08:00
Krrish Dholakia
aec4572285 fix(proxy_cli.py): fix max connection limit issue on db 2024-02-07 22:57:44 -08:00
Krrish Dholakia
c429dfa746 fix(proxy_cli.py-&&-proxy_server.py): bump reset budget intervals and fix pool limits for prisma connections 2024-02-06 19:39:49 -08:00
Krish Dholakia
45cbb3cf3d Merge branch 'main' into litellm_embedding_caching_updates 2024-02-03 18:08:47 -08:00
Krrish Dholakia
9aa55ab872 build(proxy_cli.py): make running gunicorn an optional cli arg
when running proxy locally, running with uvicorn is much better for debugging
2024-01-29 15:32:34 -08:00
Krrish Dholakia
e937662e8a build(ui/litellm_dashboard_v_2): allow app owner to create keys and view their keys 2024-01-27 13:48:35 -08:00
Krrish Dholakia
9c0179d132 fix(proxy_server.py): support setting tpm/rpm limits per user / per key 2024-01-18 17:03:18 -08:00
Krrish Dholakia
3bdfb2bc6d fix(proxy_cli.py): ensure proxy always retries if db push fails to connect to db 2024-01-17 17:37:59 -08:00
ishaan-jaff
b01321cdc7 (feat) use env variables to set host, port, num_workers 2024-01-16 08:36:29 -08:00
ishaan-jaff
1ff697af3e (feat) print num_workers when starting prox 2024-01-16 08:21:37 -08:00
ishaan-jaff
b3de8e8fa2 v0 num_workers 2024-01-16 08:15:12 -08:00
ishaan-jaff
4a7f42cb73 (fix) gunicorn - show request logs like uvicorn 2024-01-13 11:22:16 -08:00
Krrish Dholakia
61cbac6b4f fix(proxy_cli.py): only run prisma db push if prisma in environment 2024-01-13 19:44:12 +05:30
Krrish Dholakia
813fb19620 fix: support async redis caching 2024-01-12 21:46:41 +05:30
XD3000/高瑞雲
a945ffa9c7 #1424:gunicorn can't run in windows 2024-01-12 14:30:23 +08:00
Krrish Dholakia
36068b707a fix(proxy_cli.py): read db url from config, not just environment 2024-01-11 19:19:29 +05:30
Krrish Dholakia
3028d0f622 feat(proxy_cli.py): move print statements to show actually deployed port 2024-01-10 22:09:58 +05:30
Ishaan Jaff
6d994996bc Merge pull request #1399 from BerriAI/litellm_default_use_gunicorn
LiteLLM Proxy - Use Gunicorn with Uvicorn workers
2024-01-10 21:46:04 +05:30
Krrish Dholakia
e08da07def build(config.yml): reintroduce mounting config.yaml 2024-01-10 18:03:57 +05:30
ishaan-jaff
61e38e9d82 (feat) add comments on starting with gunicorn 2024-01-10 17:50:51 +05:30
ishaan-jaff
5e53ed8178 (fix) import gunicorn 2024-01-10 17:47:34 +05:30
ishaan-jaff
f310b6b7f3 (chore) remove old uvicorn logic 2024-01-10 17:39:05 +05:30
ishaan-jaff
fe122bf606 (fix) use gunicorn to start proxt 2024-01-10 17:09:03 +05:30
ishaan-jaff
ee90995272 v0 2024-01-10 16:29:38 +05:30
ishaan-jaff
729637e2ad (feat) allow users to opt into detailed debug on proxy 2024-01-08 12:53:41 +05:30
ishaan-jaff
e1ce865dbc (fix) proxy - raise error when user missing litellm[proxy] 2023-12-28 13:07:44 +05:30
ishaan-jaff
3ea4eb2fc6 (feat) proxy - remove appdirs 2023-12-27 17:40:05 +05:30
ishaan-jaff
a8468e30f0 (feat) proxy, use --model with --test 2023-12-26 09:40:58 +05:30
Krrish Dholakia
79978c44ba refactor: add black formatting 2023-12-25 14:11:20 +05:30
ishaan-jaff
6d120b7e9f (fix) proxy cli --version 2023-12-14 13:22:39 +05:30
ishaan-jaff
72fe8fea04 (feat) proxy add --version 2023-12-14 12:28:42 +05:30
Krrish Dholakia
cfab098a9c fix: fix run_ollama_serve to only run if api base is none 2023-12-09 21:31:46 -08:00
Krrish Dholakia
2048ef445d fix(proxy_server.py): fix pydantic version errors 2023-12-09 12:09:49 -08:00
ishaan-jaff
aded35d3c8 (feat) proxy cli, better description of config yaml param 2023-12-05 18:11:29 -08:00
ishaan-jaff
9693e5ede6 (fix) prox cli: remove deprecated param 2023-12-05 18:04:08 -08:00
ishaan-jaff
3fbd2a853f (feat) --health for checking config models 2023-11-27 12:13:21 -08:00
Krrish Dholakia
0e3064ac8c fix(router.py): fix caching for tracking cooldowns + usage 2023-11-23 11:13:32 -08:00
Krrish Dholakia
550ddb4a6b docs(routing.md): update routing docs 2023-11-21 19:32:50 -08:00
Krrish Dholakia
9205f70b0f docs(routing.md): add queueing to docs 2023-11-21 18:01:02 -08:00
ishaan-jaff
05f9ff6351 (fix) proxy server LiteLLM warning 2023-11-21 08:50:31 -08:00
Krrish Dholakia
7472be1529 fix(routing.py): update token usage on streaming 2023-11-20 14:19:25 -08:00
Krrish Dholakia
5d58bb9cd0 fix(main.py): misrouting ollama models to nlp cloud 2023-11-14 18:55:08 -08:00
ishaan-jaff
aec5ad0f66 (fix) proxy cli compatible with openai v1.0.0 2023-11-13 10:58:20 -08:00
ishaan-jaff
cdd40ae45f (fix) proxy cli use openai v1.0.0 2023-11-13 10:08:48 -08:00
ishaan-jaff
523fa7c81c (fix) proxy + docs: use openai.chat.completions.create instead of openai.ChatCompletions 2023-11-13 08:24:26 -08:00
ishaan-jaff
7eeb931021 (fix) proxy raise exception when config passed in 2023-11-10 16:28:34 -08:00
ishaan-jaff
0e8a034360 (fix) proxy cli default local debug to False 2023-11-09 11:30:11 -08:00