Krrish Dholakia
|
6506fba3bc
|
test(test_proxy_exception_mapping.py): fix exception checking
|
2024-01-04 22:45:16 +05:30 |
|
Krrish Dholakia
|
25241de69e
|
fix(router.py): don't retry malformed / content policy violating errors (400 status code)
https://github.com/BerriAI/litellm/issues/1317 , https://github.com/BerriAI/litellm/issues/1316
|
2024-01-04 22:23:51 +05:30 |
|
Krrish Dholakia
|
74f6f6489a
|
fix(proxy_server.py): fix prisma client connection error
|
2024-01-04 18:28:18 +05:30 |
|
Krrish Dholakia
|
c7644915f9
|
fix(test_proxy_server.py): fix import
|
2024-01-04 16:11:23 +05:30 |
|
ishaan-jaff
|
c231a6e4d3
|
(ci/cd) run proxy test with debug=True
|
2024-01-04 13:01:00 +05:30 |
|
ishaan-jaff
|
234c057e97
|
(fix) azure+cf gateway, health check
|
2024-01-04 12:34:07 +05:30 |
|
Krrish Dholakia
|
b0827a87b2
|
fix(caching.py): support s-maxage param for cache controls
|
2024-01-04 11:41:23 +05:30 |
|
ishaan-jaff
|
54653f9a4a
|
(test) proxy + s3 caching
|
2024-01-04 11:11:08 +05:30 |
|
ishaan-jaff
|
aa757d19f5
|
(test) router - init clients - azure cloudflare, openai etc
|
2024-01-04 10:55:18 +05:30 |
|
ishaan-jaff
|
0864713b62
|
(test) cf azure
|
2024-01-04 10:26:41 +05:30 |
|
ishaan-jaff
|
6d21ee3a2f
|
(fix) proxy - cloudflare + Azure bug [non-streaming]
|
2024-01-04 10:24:51 +05:30 |
|
ishaan-jaff
|
d14a41863f
|
(test) s3 cache with setting s3_bucket_name
|
2024-01-03 15:42:23 +05:30 |
|
ishaan-jaff
|
fea0a933ae
|
(test) use s3 buckets cache
|
2024-01-03 15:13:43 +05:30 |
|
Krrish Dholakia
|
8cee267a5b
|
fix(caching.py): support ttl, s-max-age, and no-cache cache controls
https://github.com/BerriAI/litellm/issues/1306
|
2024-01-03 12:42:43 +05:30 |
|
ishaan-jaff
|
2bea0c742e
|
(test) completion tokens counting + azure stream
|
2024-01-03 12:06:39 +05:30 |
|
ishaan-jaff
|
14738ec89d
|
(test) xinference on litellm router
|
2024-01-02 16:51:08 +05:30 |
|
ishaan-jaff
|
bfbed2d93d
|
(test) xinference embeddings
|
2024-01-02 15:41:51 +05:30 |
|
Krrish Dholakia
|
0fffcc1579
|
fix(utils.py): support token counting for gpt-4-vision models
|
2024-01-02 14:41:42 +05:30 |
|
ishaan-jaff
|
bfae0fe935
|
(test) proxy - pass user_config
|
2024-01-02 14:15:03 +05:30 |
|
Krrish Dholakia
|
2ab31bcaf8
|
fix(lowest_tpm_rpm.py): handle null case for text/message input
|
2024-01-02 12:24:29 +05:30 |
|
ishaan-jaff
|
31a896908b
|
(test) proxy - use, user provided model_list
|
2024-01-02 12:10:34 +05:30 |
|
Krrish Dholakia
|
a37a18ca80
|
feat(router.py): add support for retry/fallbacks for async embedding calls
|
2024-01-02 11:54:28 +05:30 |
|
Krrish Dholakia
|
dff4c172d0
|
refactor(test_router_caching.py): move tpm/rpm routing tests to separate file
|
2024-01-02 11:10:11 +05:30 |
|
ishaan-jaff
|
18ef244230
|
(test) bedrock-test passing boto3 client
|
2024-01-02 10:23:28 +05:30 |
|
Ishaan Jaff
|
9adcfedc04
|
(test) fix test_get_model_cost_map.py
|
2024-01-01 21:58:48 +05:30 |
|
Krrish Dholakia
|
a83e2e07cf
|
fix(router.py): correctly raise no model available error
https://github.com/BerriAI/litellm/issues/1289
|
2024-01-01 21:22:42 +05:30 |
|
ishaan-jaff
|
84cfa1c42a
|
(test) ci/cd
|
2024-01-01 13:51:27 +05:30 |
|
ishaan-jaff
|
84fbc903aa
|
(test) langfuse - set custom trace_id
|
2023-12-30 20:19:22 +05:30 |
|
ishaan-jaff
|
e35f17ca3c
|
(test) caching - context managers
|
2023-12-30 19:33:47 +05:30 |
|
Krrish Dholakia
|
027218c3f0
|
test(test_lowest_latency_routing.py): add more tests
|
2023-12-30 17:41:42 +05:30 |
|
Krrish Dholakia
|
f2d0d5584a
|
fix(router.py): fix latency based routing
|
2023-12-30 17:25:40 +05:30 |
|
Krrish Dholakia
|
c41b1418d4
|
test(test_router_init.py): fix test router init
|
2023-12-30 16:51:39 +05:30 |
|
Krrish Dholakia
|
3cb7acceaa
|
test(test_least_busy_routing.py): fix test
|
2023-12-30 16:12:52 +05:30 |
|
Krrish Dholakia
|
3935f99083
|
test(test_router.py): add retries
|
2023-12-30 15:54:46 +05:30 |
|
Krrish Dholakia
|
69935db239
|
fix(router.py): periodically re-initialize azure/openai clients to solve max conn issue
|
2023-12-30 15:48:34 +05:30 |
|
Krrish Dholakia
|
b66cf0aa43
|
fix(lowest_tpm_rpm_routing.py): broaden scope of get deployment logic
|
2023-12-30 13:27:50 +05:30 |
|
Krrish Dholakia
|
2acd086596
|
test(test_least_busy_routing.py): fix test init
|
2023-12-30 12:39:13 +05:30 |
|
Krrish Dholakia
|
c33c1d85bb
|
fix: support dynamic timeouts for openai and azure
|
2023-12-30 12:14:02 +05:30 |
|
Krrish Dholakia
|
77be3e3114
|
fix(main.py): don't set timeout as an optional api param
|
2023-12-30 11:47:07 +05:30 |
|
Krrish Dholakia
|
38f55249e1
|
fix(router.py): support retry and fallbacks for atext_completion
|
2023-12-30 11:19:32 +05:30 |
|
ishaan-jaff
|
523415cb0c
|
(test) dynamic timeout on router
|
2023-12-30 10:56:07 +05:30 |
|
Krrish Dholakia
|
a34de56289
|
fix(router.py): handle initial scenario for tpm/rpm routing
|
2023-12-30 07:28:45 +05:30 |
|
Krrish Dholakia
|
2fc264ca04
|
fix(router.py): fix int logic
|
2023-12-29 20:41:56 +05:30 |
|
Krrish Dholakia
|
cf91e49c87
|
refactor(lowest_tpm_rpm.py): move tpm/rpm based routing to a separate file for better testing
|
2023-12-29 18:33:43 +05:30 |
|
Krrish Dholakia
|
54d7bc2cc3
|
test(test_least_busy_router.py): add better testing for least busy routing
|
2023-12-29 17:16:00 +05:30 |
|
Krrish Dholakia
|
678bbfa9be
|
fix(least_busy.py): support consistent use of model id instead of deployment name
|
2023-12-29 17:05:26 +05:30 |
|
ishaan-jaff
|
06e4b301b4
|
(test) gemini-pro-vision cost tracking
|
2023-12-29 16:31:28 +05:30 |
|
ishaan-jaff
|
e6a7212d10
|
(fix) counting streaming prompt tokens - azure
|
2023-12-29 16:13:52 +05:30 |
|
ishaan-jaff
|
73f60b7315
|
(test) stream chunk builder - azure prompt tokens
|
2023-12-29 15:45:41 +05:30 |
|
ishaan-jaff
|
b1077ebc38
|
(test) test_token_counter_azure
|
2023-12-29 15:37:46 +05:30 |
|