Commit graph

115 commits

Author SHA1 Message Date
Krrish Dholakia
ff93609453 build(schema.prisma): support direct url on prisma schema 2024-02-08 17:37:37 -08:00
Krrish Dholakia
0d803e1379 fix(proxy_cli.py): fix max connection limit issue on db 2024-02-07 22:57:44 -08:00
Krrish Dholakia
4a0df3cb4f fix(proxy_cli.py-&&-proxy_server.py): bump reset budget intervals and fix pool limits for prisma connections 2024-02-06 19:39:49 -08:00
Krish Dholakia
9ab59045a3
Merge branch 'main' into litellm_embedding_caching_updates 2024-02-03 18:08:47 -08:00
Krrish Dholakia
8e9197b5b4 build(proxy_cli.py): make running gunicorn an optional cli arg
when running proxy locally, running with uvicorn is much better for debugging
2024-01-29 15:32:34 -08:00
Krrish Dholakia
dda115fcb7 build(ui/litellm_dashboard_v_2): allow app owner to create keys and view their keys 2024-01-27 13:48:35 -08:00
Krrish Dholakia
1e5efdfa37 fix(proxy_server.py): support setting tpm/rpm limits per user / per key 2024-01-18 17:03:18 -08:00
Krrish Dholakia
73daee7e07 fix(proxy_cli.py): ensure proxy always retries if db push fails to connect to db 2024-01-17 17:37:59 -08:00
ishaan-jaff
893627cab4 (feat) use env variables to set host, port, num_workers 2024-01-16 08:36:29 -08:00
ishaan-jaff
6a09fcfa06 (feat) print num_workers when starting prox 2024-01-16 08:21:37 -08:00
ishaan-jaff
7b1c8471c5 v0 num_workers 2024-01-16 08:15:12 -08:00
ishaan-jaff
61dff3decb (fix) gunicorn - show request logs like uvicorn 2024-01-13 11:22:16 -08:00
Krrish Dholakia
3d7c169a58 fix(proxy_cli.py): only run prisma db push if prisma in environment 2024-01-13 19:44:12 +05:30
Krrish Dholakia
007870390d fix: support async redis caching 2024-01-12 21:46:41 +05:30
XD3000/高瑞雲
574d042655 #1424:gunicorn can't run in windows 2024-01-12 14:30:23 +08:00
Krrish Dholakia
43533812a7 fix(proxy_cli.py): read db url from config, not just environment 2024-01-11 19:19:29 +05:30
Krrish Dholakia
5bc44353e0 feat(proxy_cli.py): move print statements to show actually deployed port 2024-01-10 22:09:58 +05:30
Ishaan Jaff
58d0366447
Merge pull request #1399 from BerriAI/litellm_default_use_gunicorn
LiteLLM Proxy - Use Gunicorn with Uvicorn workers
2024-01-10 21:46:04 +05:30
Krrish Dholakia
b06d7f0cb6 build(config.yml): reintroduce mounting config.yaml 2024-01-10 18:03:57 +05:30
ishaan-jaff
2b9174c3d7 (feat) add comments on starting with gunicorn 2024-01-10 17:50:51 +05:30
ishaan-jaff
67dc9adc71 (fix) import gunicorn 2024-01-10 17:47:34 +05:30
ishaan-jaff
873965df22 (chore) remove old uvicorn logic 2024-01-10 17:39:05 +05:30
ishaan-jaff
5136d5980f (fix) use gunicorn to start proxt 2024-01-10 17:09:03 +05:30
ishaan-jaff
c7fe33202d v0 2024-01-10 16:29:38 +05:30
ishaan-jaff
6786e4f343 (feat) allow users to opt into detailed debug on proxy 2024-01-08 12:53:41 +05:30
ishaan-jaff
ab90b547d8 (fix) proxy - raise error when user missing litellm[proxy] 2023-12-28 13:07:44 +05:30
ishaan-jaff
c1a8e30b01 (feat) proxy - remove appdirs 2023-12-27 17:40:05 +05:30
ishaan-jaff
d273d19bd9 (feat) proxy, use --model with --test 2023-12-26 09:40:58 +05:30
Krrish Dholakia
4905929de3 refactor: add black formatting 2023-12-25 14:11:20 +05:30
ishaan-jaff
77bcaaae9e (fix) proxy cli --version 2023-12-14 13:22:39 +05:30
ishaan-jaff
241add8b33 (feat) proxy add --version 2023-12-14 12:28:42 +05:30
Krrish Dholakia
f10bb708c0 fix: fix run_ollama_serve to only run if api base is none 2023-12-09 21:31:46 -08:00
Krrish Dholakia
ed50522863 fix(proxy_server.py): fix pydantic version errors 2023-12-09 12:09:49 -08:00
ishaan-jaff
27d7d7ba9c (feat) proxy cli, better description of config yaml param 2023-12-05 18:11:29 -08:00
ishaan-jaff
155e99b9a3 (fix) prox cli: remove deprecated param 2023-12-05 18:04:08 -08:00
ishaan-jaff
9747cc5aad (feat) --health for checking config models 2023-11-27 12:13:21 -08:00
Krrish Dholakia
61fc76a8c4 fix(router.py): fix caching for tracking cooldowns + usage 2023-11-23 11:13:32 -08:00
Krrish Dholakia
826f56a6a0 docs(routing.md): update routing docs 2023-11-21 19:32:50 -08:00
Krrish Dholakia
9d97082eed docs(routing.md): add queueing to docs 2023-11-21 18:01:02 -08:00
ishaan-jaff
2a35ff88a7 (fix) proxy server LiteLLM warning 2023-11-21 08:50:31 -08:00
Krrish Dholakia
1976d0f7d6 fix(routing.py): update token usage on streaming 2023-11-20 14:19:25 -08:00
Krrish Dholakia
1738341dcb fix(main.py): misrouting ollama models to nlp cloud 2023-11-14 18:55:08 -08:00
ishaan-jaff
e125414611 (fix) proxy cli compatible with openai v1.0.0 2023-11-13 10:58:20 -08:00
ishaan-jaff
18b694f01a (fix) proxy cli use openai v1.0.0 2023-11-13 10:08:48 -08:00
ishaan-jaff
cf0ab7155e (fix) proxy + docs: use openai.chat.completions.create instead of openai.ChatCompletions 2023-11-13 08:24:26 -08:00
ishaan-jaff
78e1ed9575 (fix) proxy raise exception when config passed in 2023-11-10 16:28:34 -08:00
ishaan-jaff
333268c8b7 (fix) proxy cli default local debug to False 2023-11-09 11:30:11 -08:00
ishaan-jaff
24c0a65347 (fix) proxy server clean print statements 2023-11-09 11:18:56 -08:00
ishaan-jaff
03940eab8a (fix) prxy server remove create_proxy 2023-11-09 11:12:20 -08:00
ishaan-jaff
285c678786 (fix) proxy remove --create_proxy 2023-11-09 11:10:08 -08:00