Commit graph

1456 commits

Author SHA1 Message Date
ishaan-jaff
6f9d3fc3bc (ci/cd) retry hosted ollama + stream test 3 times 2024-01-05 18:02:20 +05:30
ishaan-jaff
0eb899c087 (test) hosted ollama - retry 3 times 2024-01-05 17:58:59 +05:30
ishaan-jaff
76b2db4492 (ci/cd) run test again 2024-01-05 16:40:56 +05:30
ishaan-jaff
69bac0dbf6 (ci/cd) test proxy - init prisma in test 2024-01-05 16:18:23 +05:30
ishaan-jaff
4679c7b99a (fix) caching use same "created" in response_object 2024-01-05 16:03:56 +05:30
ishaan-jaff
f211009263 (test) openai embedding cost calculation 2024-01-05 15:22:17 +05:30
ishaan-jaff
20256c45ad (fix) retry cloudflare ai workers 3 times 2024-01-05 13:55:47 +05:30
ishaan-jaff
6694975ec3 (test) azure completion_cost 2024-01-05 13:53:08 +05:30
ishaan-jaff
13201edc4b (test) test reading configs on proxy 2024-01-05 13:37:31 +05:30
ishaan-jaff
72e7178c9b (test) azure/embedding + completion_cost 2024-01-05 13:19:17 +05:30
ishaan-jaff
f681f0f2b2 (feat) completion_cost - embeddings + raise Exception 2024-01-05 13:11:23 +05:30
ishaan-jaff
113b5e7284 (ci/cd) retry cloudflare request 3 times 2024-01-05 12:40:53 +05:30
ishaan-jaff
83b31141c6 (ci/cd) raise correct exception proxy 2024-01-05 12:29:03 +05:30
ishaan-jaff
bcf22725a6 (ci/cd) run cloudflare test 3 retries 2024-01-05 11:55:12 +05:30
ishaan-jaff
d1865591aa (fix) test caching- use azure, instead of bedrock 2024-01-05 10:51:56 +05:30
Krrish Dholakia
6506fba3bc test(test_proxy_exception_mapping.py): fix exception checking 2024-01-04 22:45:16 +05:30
Krrish Dholakia
25241de69e fix(router.py): don't retry malformed / content policy violating errors (400 status code)
https://github.com/BerriAI/litellm/issues/1317 , https://github.com/BerriAI/litellm/issues/1316
2024-01-04 22:23:51 +05:30
Krrish Dholakia
74f6f6489a fix(proxy_server.py): fix prisma client connection error 2024-01-04 18:28:18 +05:30
Krrish Dholakia
c7644915f9 fix(test_proxy_server.py): fix import 2024-01-04 16:11:23 +05:30
ishaan-jaff
c231a6e4d3 (ci/cd) run proxy test with debug=True 2024-01-04 13:01:00 +05:30
ishaan-jaff
234c057e97 (fix) azure+cf gateway, health check 2024-01-04 12:34:07 +05:30
Krrish Dholakia
b0827a87b2 fix(caching.py): support s-maxage param for cache controls 2024-01-04 11:41:23 +05:30
ishaan-jaff
54653f9a4a (test) proxy + s3 caching 2024-01-04 11:11:08 +05:30
ishaan-jaff
aa757d19f5 (test) router - init clients - azure cloudflare, openai etc 2024-01-04 10:55:18 +05:30
ishaan-jaff
0864713b62 (test) cf azure 2024-01-04 10:26:41 +05:30
ishaan-jaff
6d21ee3a2f (fix) proxy - cloudflare + Azure bug [non-streaming] 2024-01-04 10:24:51 +05:30
ishaan-jaff
d14a41863f (test) s3 cache with setting s3_bucket_name 2024-01-03 15:42:23 +05:30
ishaan-jaff
fea0a933ae (test) use s3 buckets cache 2024-01-03 15:13:43 +05:30
Krrish Dholakia
8cee267a5b fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
2024-01-03 12:42:43 +05:30
ishaan-jaff
2bea0c742e (test) completion tokens counting + azure stream 2024-01-03 12:06:39 +05:30
ishaan-jaff
14738ec89d (test) xinference on litellm router 2024-01-02 16:51:08 +05:30
ishaan-jaff
bfbed2d93d (test) xinference embeddings 2024-01-02 15:41:51 +05:30
Krrish Dholakia
0fffcc1579 fix(utils.py): support token counting for gpt-4-vision models 2024-01-02 14:41:42 +05:30
ishaan-jaff
bfae0fe935 (test) proxy - pass user_config 2024-01-02 14:15:03 +05:30
Krrish Dholakia
2ab31bcaf8 fix(lowest_tpm_rpm.py): handle null case for text/message input 2024-01-02 12:24:29 +05:30
ishaan-jaff
31a896908b (test) proxy - use, user provided model_list 2024-01-02 12:10:34 +05:30
Krrish Dholakia
a37a18ca80 feat(router.py): add support for retry/fallbacks for async embedding calls 2024-01-02 11:54:28 +05:30
Krrish Dholakia
dff4c172d0 refactor(test_router_caching.py): move tpm/rpm routing tests to separate file 2024-01-02 11:10:11 +05:30
ishaan-jaff
18ef244230 (test) bedrock-test passing boto3 client 2024-01-02 10:23:28 +05:30
Ishaan Jaff
9adcfedc04
(test) fix test_get_model_cost_map.py 2024-01-01 21:58:48 +05:30
Krrish Dholakia
a83e2e07cf fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
2024-01-01 21:22:42 +05:30
ishaan-jaff
84cfa1c42a (test) ci/cd 2024-01-01 13:51:27 +05:30
ishaan-jaff
84fbc903aa (test) langfuse - set custom trace_id 2023-12-30 20:19:22 +05:30
ishaan-jaff
e35f17ca3c (test) caching - context managers 2023-12-30 19:33:47 +05:30
Krrish Dholakia
027218c3f0 test(test_lowest_latency_routing.py): add more tests 2023-12-30 17:41:42 +05:30
Krrish Dholakia
f2d0d5584a fix(router.py): fix latency based routing 2023-12-30 17:25:40 +05:30
Krrish Dholakia
c41b1418d4 test(test_router_init.py): fix test router init 2023-12-30 16:51:39 +05:30
Krrish Dholakia
3cb7acceaa test(test_least_busy_routing.py): fix test 2023-12-30 16:12:52 +05:30
Krrish Dholakia
3935f99083 test(test_router.py): add retries 2023-12-30 15:54:46 +05:30
Krrish Dholakia
69935db239 fix(router.py): periodically re-initialize azure/openai clients to solve max conn issue 2023-12-30 15:48:34 +05:30