Commit graph

5762 commits

Author SHA1 Message Date
ishaan-jaff
872a36b7e5 (fix) minimize dockerfile changes 2024-01-09 14:32:09 +05:30
ishaan-jaff
98d8dc3331 (debug) always do prisma_setup after intiialize 2024-01-09 14:26:30 +05:30
ishaan-jaff
8ac0455ef7 test conditional prisma generate 2024-01-09 13:51:50 +05:30
ishaan-jaff
28f7ccbcd1 (temp) use database dockerfile for railway 2024-01-09 13:44:37 +05:30
Krrish Dholakia
a4e51362b5 docs(streaming_logging.md): fix tutorial 2024-01-09 13:40:37 +05:30
ishaan-jaff
599318b40e raise exception when prisma init fails 2024-01-09 13:36:02 +05:30
Krrish Dholakia
d6361e155a docs(deploy.md): update docs for deploying proxy with connected db 2024-01-09 13:34:47 +05:30
ishaan-jaff
8f8237a1a0 (fix) echo DB URL 2024-01-09 13:30:49 +05:30
ishaan-jaff
358a44ce13 (fix) dockerfile run prisma generate if DATABSE_URL set in env 2024-01-09 13:10:03 +05:30
ishaan-jaff
46bd99ad98 (test) test deployed proxy keygen 2024-01-09 13:03:22 +05:30
Krish Dholakia
5da88809b7
Update README.md 2024-01-09 12:53:01 +05:30
Krish Dholakia
14b377d01a
Update README.md 2024-01-09 12:51:30 +05:30
Krish Dholakia
59505eae03
Update README.md 2024-01-09 12:50:51 +05:30
Krish Dholakia
d916527667
Update README.md 2024-01-09 12:49:17 +05:30
Krish Dholakia
de816f4732
Update README.md 2024-01-09 12:48:49 +05:30
Krish Dholakia
b07a738652
Update README.md 2024-01-09 12:48:11 +05:30
Krish Dholakia
8dfbdc521b
Update README.md 2024-01-09 12:47:08 +05:30
Krrish Dholakia
9d45c73475 build(model_prices_and_context_window.json): add max input tokens for openai and azure models 2024-01-09 12:23:57 +05:30
Krrish Dholakia
2f377db604 build(model_prices_and_context_window.json): add max output tokens for openai + azure models, remove shutdown openai models 2024-01-09 12:16:02 +05:30
Krrish Dholakia
ed6ae8600f fix(openai.py): fix exception raising logic 2024-01-09 11:58:30 +05:30
Krrish Dholakia
be1e101b5f fix(azure.py,-openai.py): raise the correct exceptions for image generation calls 2024-01-09 11:55:38 +05:30
Krrish Dholakia
cd350ab8d8 fix(proxy_server.py): don't reconnect prisma if already connected 2024-01-09 11:38:42 +05:30
Krrish Dholakia
e97eff4243 test(test_router.py): fix router test 2024-01-09 11:08:35 +05:30
ishaan-jaff
f46fa2b8a8 (fix) test - deprecated textdavinci003 2024-01-09 10:55:35 +05:30
ishaan-jaff
9c7a4fde87 (test) hosted - ollama catch timeouts 2024-01-09 10:35:29 +05:30
ishaan-jaff
5f2cbfc711 (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
e99a41307a test: testing fixes 2024-01-09 10:23:34 +05:30
ishaan-jaff
08525ce200 (ci/cd) use 3 retries for image generation 2024-01-09 10:07:09 +05:30
ishaan-jaff
9be7e34cb0 (ci/cd) pytest skip slow replicate test 2024-01-09 09:57:06 +05:30
Krish Dholakia
e483df014b
Merge pull request #1369 from BerriAI/litellm_lowest_latency_routing
feat(lowest_latency.py): support expanded time window for latency based routing
2024-01-09 09:52:39 +05:30
Krrish Dholakia
a35f4272f4 refactor(lowest_latency.py): fix linting error 2024-01-09 09:51:43 +05:30
Krrish Dholakia
88d498a54a fix(ollama.py): use tiktoken as backup for prompt token counting 2024-01-09 09:47:18 +05:30
Krrish Dholakia
11b6c66609 docs(gemini.md): fix docs 2024-01-09 09:38:04 +05:30
Krrish Dholakia
a5147f9e06 feat(lowest_latency.py): support expanded time window for latency based routing
uses a 1hr avg. of latency for deployments, to determine which to route to

https://github.com/BerriAI/litellm/issues/1361
2024-01-09 09:38:04 +05:30
Krish Dholakia
8225eda0cc
Merge pull request #1370 from haseeb-heaven/main
Updated Gemini AI Documentation
2024-01-09 09:36:54 +05:30
HeavenHM
fa9d0faf13
Update gemini.md
Added example for Gemini Vision Pro
2024-01-09 08:00:51 +05:30
Mateo Cámara
3bc8a03150 Merge remote-tracking branch 'origin/main'
# Conflicts:
#	litellm/main.py
2024-01-08 18:51:46 +01:00
Krrish Dholakia
5b7c3c7187 refactor(lowest_latency.py): fix linting issue 2024-01-08 23:07:43 +05:30
Krrish Dholakia
a60e23d98a feat(lowest_latency.py): support expanded time window for latency based routing
uses a 1hr avg. of latency for deployments, to determine which to route to

https://github.com/BerriAI/litellm/issues/1361
2024-01-08 22:52:32 +05:30
Ishaan Jaff
8a85b719f7
Merge pull request #1368 from deepinfra/udpate-models-2
Update deepinfra models
2024-01-08 22:46:14 +05:30
ishaan-jaff
6263103680 (ci/cd) run again 2024-01-08 22:42:31 +05:30
Iskren Chernev
2486f92523 Update deepinfra models 2024-01-08 18:54:15 +02:00
Krrish Dholakia
edc088f038 build(Dockerfile): pip install from wheels not re-install requirements.txt
reduce size of dockerbuild

n
2024-01-08 20:26:09 +05:30
Krrish Dholakia
8edd3fe651 test(test_proxy_startup.py): fix gunicorn test 2024-01-08 19:55:18 +05:30
Krish Dholakia
59c57f84cf
Update README.md 2024-01-08 19:49:43 +05:30
Krrish Dholakia
55e70aa93a bump: version 1.16.19 → 1.16.20 2024-01-08 19:47:10 +05:30
Krish Dholakia
e949a2ada3
Merge pull request #1367 from BerriAI/litellm_proxy_startup
fix(proxy_server.py): add support for passing in config file via worker_config directly + testing
2024-01-08 19:46:48 +05:30
Krrish Dholakia
750330509e build(Dockerfile.database): fix new dockerfile 2024-01-08 19:37:34 +05:30
Krrish Dholakia
4ff4180a53 build(Dockerfile.database): fixing build issues 2024-01-08 19:31:19 +05:30
Krish Dholakia
6b3cf217a4
Update ghcr_deploy.yml 2024-01-08 18:16:46 +05:30