Krrish Dholakia
|
b66cf0aa43
|
fix(lowest_tpm_rpm_routing.py): broaden scope of get deployment logic
|
2023-12-30 13:27:50 +05:30 |
|
Krrish Dholakia
|
2acd086596
|
test(test_least_busy_routing.py): fix test init
|
2023-12-30 12:39:13 +05:30 |
|
Krrish Dholakia
|
c33c1d85bb
|
fix: support dynamic timeouts for openai and azure
|
2023-12-30 12:14:02 +05:30 |
|
Krrish Dholakia
|
77be3e3114
|
fix(main.py): don't set timeout as an optional api param
|
2023-12-30 11:47:07 +05:30 |
|
Krrish Dholakia
|
38f55249e1
|
fix(router.py): support retry and fallbacks for atext_completion
|
2023-12-30 11:19:32 +05:30 |
|
ishaan-jaff
|
523415cb0c
|
(test) dynamic timeout on router
|
2023-12-30 10:56:07 +05:30 |
|
Krrish Dholakia
|
a34de56289
|
fix(router.py): handle initial scenario for tpm/rpm routing
|
2023-12-30 07:28:45 +05:30 |
|
Krrish Dholakia
|
2fc264ca04
|
fix(router.py): fix int logic
|
2023-12-29 20:41:56 +05:30 |
|
Krrish Dholakia
|
cf91e49c87
|
refactor(lowest_tpm_rpm.py): move tpm/rpm based routing to a separate file for better testing
|
2023-12-29 18:33:43 +05:30 |
|
Krrish Dholakia
|
54d7bc2cc3
|
test(test_least_busy_router.py): add better testing for least busy routing
|
2023-12-29 17:16:00 +05:30 |
|
Krrish Dholakia
|
678bbfa9be
|
fix(least_busy.py): support consistent use of model id instead of deployment name
|
2023-12-29 17:05:26 +05:30 |
|
ishaan-jaff
|
06e4b301b4
|
(test) gemini-pro-vision cost tracking
|
2023-12-29 16:31:28 +05:30 |
|
ishaan-jaff
|
e6a7212d10
|
(fix) counting streaming prompt tokens - azure
|
2023-12-29 16:13:52 +05:30 |
|
ishaan-jaff
|
73f60b7315
|
(test) stream chunk builder - azure prompt tokens
|
2023-12-29 15:45:41 +05:30 |
|
ishaan-jaff
|
b1077ebc38
|
(test) test_token_counter_azure
|
2023-12-29 15:37:46 +05:30 |
|
ishaan-jaff
|
4a028d012a
|
(test) token_counter - prompt tokens == tokens from API
|
2023-12-29 15:15:39 +05:30 |
|
Krrish Dholakia
|
1e07f0fce8
|
fix(caching.py): hash the cache key to prevent key too long errors
|
2023-12-29 15:03:33 +05:30 |
|
Krrish Dholakia
|
6e68cd1125
|
docs(load_test.md): add litellm load test script to docs
|
2023-12-29 13:41:44 +05:30 |
|
ishaan-jaff
|
243ad31e90
|
(test) async + stream clooudflare
|
2023-12-29 12:03:29 +05:30 |
|
ishaan-jaff
|
a999e80b46
|
(test) async cloudflare
|
2023-12-29 11:50:09 +05:30 |
|
Krrish Dholakia
|
e06840b571
|
refactor: move async text completion testing to test_text_completion.py
|
2023-12-29 11:46:40 +05:30 |
|
ishaan-jaff
|
5fc9524a46
|
(test) test cloudflare completion
|
2023-12-29 11:34:58 +05:30 |
|
Krrish Dholakia
|
6f2734100f
|
fix(main.py): fix async text completion streaming + add new tests
|
2023-12-29 11:33:42 +05:30 |
|
ishaan-jaff
|
2b8e2bd937
|
(ci/cd) set num retries for HF test
|
2023-12-29 10:52:45 +05:30 |
|
Krrish Dholakia
|
86403cd14e
|
fix(vertex_ai.py): support function calling for gemini
|
2023-12-28 19:07:04 +05:30 |
|
ishaan-jaff
|
95e6d2fbba
|
(feat) add voyage ai embeddings
|
2023-12-28 17:10:15 +05:30 |
|
ishaan-jaff
|
41d5dceaec
|
(test) mistral-embed
|
2023-12-28 16:42:36 +05:30 |
|
Krrish Dholakia
|
4db138facc
|
test(test_proxy_custom_logger.py): fix testing to handle [done] chunks
|
2023-12-28 11:37:57 +05:30 |
|
Krrish Dholakia
|
c4fc28ab0d
|
fix(utils.py): use local tiktoken copy
|
2023-12-28 11:22:33 +05:30 |
|
ishaan-jaff
|
7751b83100
|
(ci/cd) run render deploy
|
2023-12-28 11:16:58 +05:30 |
|
Krrish Dholakia
|
5089753925
|
test(test_health_check.py): fix azure embedding health check test
|
2023-12-28 10:48:31 +05:30 |
|
Krrish Dholakia
|
3b1685e7c6
|
feat(health_check.py): more detailed health check calls
|
2023-12-28 09:12:57 +05:30 |
|
ishaan-jaff
|
b071157df5
|
(test) proxy - debug /health
|
2023-12-27 19:18:41 +05:30 |
|
ishaan-jaff
|
ca8b87915f
|
(test) langfuse - mark beta test
|
2023-12-27 18:43:49 +05:30 |
|
ishaan-jaff
|
a16411cf1d
|
(ci/cd) move langfuse test early
|
2023-12-27 17:50:44 +05:30 |
|
ishaan-jaff
|
d7646669f2
|
(test) langfuse logging
|
2023-12-27 17:50:05 +05:30 |
|
Krrish Dholakia
|
e516cfe9f5
|
fix(utils.py): allow text completion input to be either model or engine
|
2023-12-27 17:24:16 +05:30 |
|
Krrish Dholakia
|
10a74d02c1
|
test(test_router_fallbacks.py): fix test to check previous models in pre_api_call not on success
|
2023-12-27 16:34:40 +05:30 |
|
ishaan-jaff
|
f6f8d1a9df
|
(test) langfuse beta test - text_completion
|
2023-12-27 15:45:40 +05:30 |
|
ishaan-jaff
|
8528d9f809
|
(test) gpt-3.5-turbo-instruct finish reason
|
2023-12-27 15:45:40 +05:30 |
|
Krrish Dholakia
|
31148922b3
|
fix(azure.py): raise streaming exceptions
|
2023-12-27 15:43:13 +05:30 |
|
ishaan-jaff
|
db135aea4c
|
(test) fix langfuse test
|
2023-12-27 15:20:26 +05:30 |
|
Krrish Dholakia
|
6d63c0015b
|
test(test_router_fallbacks.py): add testing for sync streaming fallbacks
|
2023-12-27 15:10:43 +05:30 |
|
Krrish Dholakia
|
c9fdbaf898
|
fix(azure.py,-openai.py): correctly raise errors if streaming calls fail
|
2023-12-27 15:08:37 +05:30 |
|
ishaan-jaff
|
db90ccd19f
|
(test) local ollama_chat
|
2023-12-26 20:11:14 +05:30 |
|
ishaan-jaff
|
0b8d9d177b
|
(test) ollama_chat acompletion without stream
|
2023-12-26 20:01:51 +05:30 |
|
Krrish Dholakia
|
f0b6b9dce2
|
fix(main.py): support ttl being set for completion, embedding, image generation calls
|
2023-12-26 17:22:40 +05:30 |
|
ishaan-jaff
|
e29dcf595e
|
(test) azure gpt-vision
|
2023-12-26 15:24:20 +05:30 |
|
ishaan-jaff
|
5643658dac
|
(test) stream + logprobs openai
|
2023-12-26 15:15:37 +05:30 |
|
ishaan-jaff
|
b9f2262d5f
|
(test) openai logprobs
|
2023-12-26 14:00:42 +05:30 |
|