ishaan-jaff
|
6f9d3fc3bc
|
(ci/cd) retry hosted ollama + stream test 3 times
|
2024-01-05 18:02:20 +05:30 |
|
ishaan-jaff
|
0eb899c087
|
(test) hosted ollama - retry 3 times
|
2024-01-05 17:58:59 +05:30 |
|
ishaan-jaff
|
76b2db4492
|
(ci/cd) run test again
|
2024-01-05 16:40:56 +05:30 |
|
ishaan-jaff
|
69bac0dbf6
|
(ci/cd) test proxy - init prisma in test
|
2024-01-05 16:18:23 +05:30 |
|
ishaan-jaff
|
4679c7b99a
|
(fix) caching use same "created" in response_object
|
2024-01-05 16:03:56 +05:30 |
|
ishaan-jaff
|
f211009263
|
(test) openai embedding cost calculation
|
2024-01-05 15:22:17 +05:30 |
|
ishaan-jaff
|
20256c45ad
|
(fix) retry cloudflare ai workers 3 times
|
2024-01-05 13:55:47 +05:30 |
|
ishaan-jaff
|
6694975ec3
|
(test) azure completion_cost
|
2024-01-05 13:53:08 +05:30 |
|
ishaan-jaff
|
13201edc4b
|
(test) test reading configs on proxy
|
2024-01-05 13:37:31 +05:30 |
|
ishaan-jaff
|
72e7178c9b
|
(test) azure/embedding + completion_cost
|
2024-01-05 13:19:17 +05:30 |
|
ishaan-jaff
|
f681f0f2b2
|
(feat) completion_cost - embeddings + raise Exception
|
2024-01-05 13:11:23 +05:30 |
|
ishaan-jaff
|
113b5e7284
|
(ci/cd) retry cloudflare request 3 times
|
2024-01-05 12:40:53 +05:30 |
|
ishaan-jaff
|
83b31141c6
|
(ci/cd) raise correct exception proxy
|
2024-01-05 12:29:03 +05:30 |
|
ishaan-jaff
|
bcf22725a6
|
(ci/cd) run cloudflare test 3 retries
|
2024-01-05 11:55:12 +05:30 |
|
ishaan-jaff
|
d1865591aa
|
(fix) test caching- use azure, instead of bedrock
|
2024-01-05 10:51:56 +05:30 |
|
Krrish Dholakia
|
6506fba3bc
|
test(test_proxy_exception_mapping.py): fix exception checking
|
2024-01-04 22:45:16 +05:30 |
|
Krrish Dholakia
|
25241de69e
|
fix(router.py): don't retry malformed / content policy violating errors (400 status code)
https://github.com/BerriAI/litellm/issues/1317 , https://github.com/BerriAI/litellm/issues/1316
|
2024-01-04 22:23:51 +05:30 |
|
Krrish Dholakia
|
74f6f6489a
|
fix(proxy_server.py): fix prisma client connection error
|
2024-01-04 18:28:18 +05:30 |
|
Krrish Dholakia
|
c7644915f9
|
fix(test_proxy_server.py): fix import
|
2024-01-04 16:11:23 +05:30 |
|
ishaan-jaff
|
c231a6e4d3
|
(ci/cd) run proxy test with debug=True
|
2024-01-04 13:01:00 +05:30 |
|
ishaan-jaff
|
234c057e97
|
(fix) azure+cf gateway, health check
|
2024-01-04 12:34:07 +05:30 |
|
Krrish Dholakia
|
b0827a87b2
|
fix(caching.py): support s-maxage param for cache controls
|
2024-01-04 11:41:23 +05:30 |
|
ishaan-jaff
|
54653f9a4a
|
(test) proxy + s3 caching
|
2024-01-04 11:11:08 +05:30 |
|
ishaan-jaff
|
aa757d19f5
|
(test) router - init clients - azure cloudflare, openai etc
|
2024-01-04 10:55:18 +05:30 |
|
ishaan-jaff
|
0864713b62
|
(test) cf azure
|
2024-01-04 10:26:41 +05:30 |
|
ishaan-jaff
|
6d21ee3a2f
|
(fix) proxy - cloudflare + Azure bug [non-streaming]
|
2024-01-04 10:24:51 +05:30 |
|
ishaan-jaff
|
d14a41863f
|
(test) s3 cache with setting s3_bucket_name
|
2024-01-03 15:42:23 +05:30 |
|
ishaan-jaff
|
fea0a933ae
|
(test) use s3 buckets cache
|
2024-01-03 15:13:43 +05:30 |
|
Krrish Dholakia
|
8cee267a5b
|
fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
|
2024-01-03 12:42:43 +05:30 |
|
ishaan-jaff
|
2bea0c742e
|
(test) completion tokens counting + azure stream
|
2024-01-03 12:06:39 +05:30 |
|
ishaan-jaff
|
14738ec89d
|
(test) xinference on litellm router
|
2024-01-02 16:51:08 +05:30 |
|
ishaan-jaff
|
bfbed2d93d
|
(test) xinference embeddings
|
2024-01-02 15:41:51 +05:30 |
|
Krrish Dholakia
|
0fffcc1579
|
fix(utils.py): support token counting for gpt-4-vision models
|
2024-01-02 14:41:42 +05:30 |
|
ishaan-jaff
|
bfae0fe935
|
(test) proxy - pass user_config
|
2024-01-02 14:15:03 +05:30 |
|
Krrish Dholakia
|
2ab31bcaf8
|
fix(lowest_tpm_rpm.py): handle null case for text/message input
|
2024-01-02 12:24:29 +05:30 |
|
ishaan-jaff
|
31a896908b
|
(test) proxy - use, user provided model_list
|
2024-01-02 12:10:34 +05:30 |
|
Krrish Dholakia
|
a37a18ca80
|
feat(router.py): add support for retry/fallbacks for async embedding calls
|
2024-01-02 11:54:28 +05:30 |
|
Krrish Dholakia
|
dff4c172d0
|
refactor(test_router_caching.py): move tpm/rpm routing tests to separate file
|
2024-01-02 11:10:11 +05:30 |
|
ishaan-jaff
|
18ef244230
|
(test) bedrock-test passing boto3 client
|
2024-01-02 10:23:28 +05:30 |
|
Ishaan Jaff
|
9adcfedc04
|
(test) fix test_get_model_cost_map.py
|
2024-01-01 21:58:48 +05:30 |
|
Krrish Dholakia
|
a83e2e07cf
|
fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
|
2024-01-01 21:22:42 +05:30 |
|
ishaan-jaff
|
84cfa1c42a
|
(test) ci/cd
|
2024-01-01 13:51:27 +05:30 |
|
ishaan-jaff
|
84fbc903aa
|
(test) langfuse - set custom trace_id
|
2023-12-30 20:19:22 +05:30 |
|
ishaan-jaff
|
e35f17ca3c
|
(test) caching - context managers
|
2023-12-30 19:33:47 +05:30 |
|
Krrish Dholakia
|
027218c3f0
|
test(test_lowest_latency_routing.py): add more tests
|
2023-12-30 17:41:42 +05:30 |
|
Krrish Dholakia
|
f2d0d5584a
|
fix(router.py): fix latency based routing
|
2023-12-30 17:25:40 +05:30 |
|
Krrish Dholakia
|
c41b1418d4
|
test(test_router_init.py): fix test router init
|
2023-12-30 16:51:39 +05:30 |
|
Krrish Dholakia
|
3cb7acceaa
|
test(test_least_busy_routing.py): fix test
|
2023-12-30 16:12:52 +05:30 |
|
Krrish Dholakia
|
3935f99083
|
test(test_router.py): add retries
|
2023-12-30 15:54:46 +05:30 |
|
Krrish Dholakia
|
69935db239
|
fix(router.py): periodically re-initialize azure/openai clients to solve max conn issue
|
2023-12-30 15:48:34 +05:30 |
|