Commit graph

3262 commits

Author SHA1 Message Date
Mateo Cámara
203089e6c7
Merge branch 'main' into explicit-args-acomplete 2024-01-09 13:07:37 +01:00
Mateo Cámara
9aedd4e794 Moved test to a new file 2024-01-09 13:02:12 +01:00
Mateo Cámara
0ec976b3d1 Reverted changes made by the IDE automatically 2024-01-09 12:55:12 +01:00
Ishaan Jaff
4cfa010dbd
Merge pull request #1381 from BerriAI/litellm_content_policy_violation_exception
[Feat] Add litellm.ContentPolicyViolationError
2024-01-09 17:18:29 +05:30
ishaan-jaff
248e5f3d92 (chore) remove deprecated completion_with_config() tests 2024-01-09 17:13:06 +05:30
ishaan-jaff
c0b56b6575 (test) catch litellm.ContentPolicyViolationError 2024-01-09 17:04:04 +05:30
ishaan-jaff
186fc4614d (feat) add ContentPolicyViolationError for azure 2024-01-09 16:58:09 +05:30
ishaan-jaff
170ae74118 (feat) add exception mapping for litellm.image_generation 2024-01-09 16:54:47 +05:30
ishaan-jaff
f0c10377cf (test) ContentPolicyViolationError 2024-01-09 16:53:57 +05:30
ishaan-jaff
9da61bdf31 (fix) ContentPolicyViolationError 2024-01-09 16:53:15 +05:30
Mateo Cámara
bb06c51ede Added test to check if acompletion is using the same parameters as CompletionRequest attributes. Added functools to client decorator to expose acompletion parameters from outside. 2024-01-09 12:06:49 +01:00
Mateo Cámara
48b2f69c93 Added the new acompletion parameters based on CompletionRequest attributes 2024-01-09 12:05:31 +01:00
ishaan-jaff
cf98343eb5 (test) content policy violation error 2024-01-09 16:34:20 +05:30
Mateo Cámara
178a57492b Merge remote-tracking branch 'origin/main' into explicit-args-acomplete 2024-01-09 12:04:04 +01:00
ishaan-jaff
09874cc83f (v0) add ContentPolicyViolationError 2024-01-09 16:33:03 +05:30
ishaan-jaff
47f56f0d19 (test) deployed proxy key/gen 2024-01-09 16:07:33 +05:30
Ishaan Jaff
cdeb864e28
Merge pull request #1376 from BerriAI/litellm_deployed_proxy_prisma
[Test+Fix] Use deployed proxy with Prisma
2024-01-09 16:02:27 +05:30
ishaan-jaff
b5f9f05491 (test) fix - skip HF is currently loading exception 2024-01-09 15:53:19 +05:30
ishaan-jaff
bf65306ec3 (chore) undo extra space 2024-01-09 15:38:14 +05:30
ishaan-jaff
0434ee4f02 (ci/cd) run again 2024-01-09 15:32:10 +05:30
ishaan-jaff
916e93f398 (ci/cd) user DATABASE_URL to control prisma generate 2024-01-09 15:28:37 +05:30
ishaan-jaff
d24d9cb673 (ci/cd) run again 2024-01-09 15:22:28 +05:30
ishaan-jaff
7e15802388 (ci/cd) run again 2024-01-09 15:14:51 +05:30
ishaan-jaff
ba4646dbb6 (ci/cd) trigger again 2024-01-09 15:06:55 +05:30
ishaan-jaff
ebb5490d3c (fix) prisma setup 2024-01-09 14:56:43 +05:30
ishaan-jaff
98d8dc3331 (debug) always do prisma_setup after intiialize 2024-01-09 14:26:30 +05:30
ishaan-jaff
599318b40e raise exception when prisma init fails 2024-01-09 13:36:02 +05:30
ishaan-jaff
8f8237a1a0 (fix) echo DB URL 2024-01-09 13:30:49 +05:30
ishaan-jaff
46bd99ad98 (test) test deployed proxy keygen 2024-01-09 13:03:22 +05:30
Krrish Dholakia
ed6ae8600f fix(openai.py): fix exception raising logic 2024-01-09 11:58:30 +05:30
Krrish Dholakia
be1e101b5f fix(azure.py,-openai.py): raise the correct exceptions for image generation calls 2024-01-09 11:55:38 +05:30
Krrish Dholakia
cd350ab8d8 fix(proxy_server.py): don't reconnect prisma if already connected 2024-01-09 11:38:42 +05:30
Krrish Dholakia
e97eff4243 test(test_router.py): fix router test 2024-01-09 11:08:35 +05:30
ishaan-jaff
f46fa2b8a8 (fix) test - deprecated textdavinci003 2024-01-09 10:55:35 +05:30
ishaan-jaff
9c7a4fde87 (test) hosted - ollama catch timeouts 2024-01-09 10:35:29 +05:30
ishaan-jaff
5f2cbfc711 (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
e99a41307a test: testing fixes 2024-01-09 10:23:34 +05:30
ishaan-jaff
08525ce200 (ci/cd) use 3 retries for image generation 2024-01-09 10:07:09 +05:30
ishaan-jaff
9be7e34cb0 (ci/cd) pytest skip slow replicate test 2024-01-09 09:57:06 +05:30
Krrish Dholakia
a35f4272f4 refactor(lowest_latency.py): fix linting error 2024-01-09 09:51:43 +05:30
Krrish Dholakia
88d498a54a fix(ollama.py): use tiktoken as backup for prompt token counting 2024-01-09 09:47:18 +05:30
Krrish Dholakia
a5147f9e06 feat(lowest_latency.py): support expanded time window for latency based routing
uses a 1hr avg. of latency for deployments, to determine which to route to

https://github.com/BerriAI/litellm/issues/1361
2024-01-09 09:38:04 +05:30
Mateo Cámara
3bc8a03150 Merge remote-tracking branch 'origin/main'
# Conflicts:
#	litellm/main.py
2024-01-08 18:51:46 +01:00
ishaan-jaff
6263103680 (ci/cd) run again 2024-01-08 22:42:31 +05:30
Krrish Dholakia
8edd3fe651 test(test_proxy_startup.py): fix gunicorn test 2024-01-08 19:55:18 +05:30
Krish Dholakia
e949a2ada3
Merge pull request #1367 from BerriAI/litellm_proxy_startup
fix(proxy_server.py): add support for passing in config file via worker_config directly + testing
2024-01-08 19:46:48 +05:30
Krrish Dholakia
e305dcf0a6 test(test_proxy_startup.py): separate tests 2024-01-08 17:58:37 +05:30
Krrish Dholakia
2bcfe28ee9 fix(proxy_server.py): improve /health/readiness endpoint to give more details on connected services 2024-01-08 17:45:00 +05:30
Krrish Dholakia
8f8f961941 fix(proxy_server.py): add support for passing in config file via worker_config directly + testing 2024-01-08 16:47:15 +05:30
ishaan-jaff
fa74831d79 (docs) control proxy debug using env vars 2024-01-08 16:15:33 +05:30