Commit graph

1911 commits

Author SHA1 Message Date
Ishaan Jaff
7cb69c72c8
Merge branch 'main' into litellm_add_semantic_cache 2024-02-06 11:18:43 -08:00
ishaan-jaff
8175fb4deb (fix) mark semantic caching as beta test 2024-02-06 11:04:19 -08:00
ishaan-jaff
405a44727c (ci/cd) run in verbose mode 2024-02-06 10:57:20 -08:00
ishaan-jaff
1afdf5cf36 (fix) semantic caching 2024-02-06 10:55:15 -08:00
ishaan-jaff
c8a83bb745 (fix) test-semantic caching 2024-02-06 10:39:44 -08:00
ishaan-jaff
a125ffe190 (test) async semantic cache 2024-02-06 08:14:54 -08:00
ishaan-jaff
47bed68c7f (fix) test_normal_router_tpm_limit 2024-02-06 06:46:49 -08:00
ishaan-jaff
9a8abdb1ae (ci/cd) print debug info for test_proxy_gunicorn_startup_config_dict 2024-02-05 22:53:31 -08:00
ishaan-jaff
4d625818d6 (fix) proxy startup test 2024-02-05 22:51:11 -08:00
ishaan-jaff
4d4554b0e4 (test) test_upperbound_key_params 2024-02-05 22:39:36 -08:00
Krrish Dholakia
7a0bccf4d0 test(test_key_generate_dynamodb.py): fix test 2024-02-05 21:44:50 -08:00
Krrish Dholakia
a9a4f4cf0f test(test_key_generate_dynamodb.py): fix test 2024-02-05 21:43:17 -08:00
ishaan-jaff
81f8ac00b2 (test) semantic caching 2024-02-05 18:22:50 -08:00
ishaan-jaff
cf4bd1cf4e (test) semantic cache 2024-02-05 17:58:32 -08:00
ishaan-jaff
2b588a8786 (test) litellm-dashboard never allowed to /chat/completions 2024-02-05 16:11:33 -08:00
ishaan-jaff
d4a799a3ca (feat )add semantic cache 2024-02-05 12:28:21 -08:00
Krish Dholakia
640572647a
Merge pull request #1805 from BerriAI/litellm_cost_tracking_image_gen
feat(utils.py): support cost tracking for openai/azure image gen models
2024-02-03 22:23:22 -08:00
Krrish Dholakia
49b2dc4180 test(test_completion_cost.py): fix test 2024-02-03 22:00:49 -08:00
Krrish Dholakia
66565f96b1 test(test_completion.py): fix test 2024-02-03 21:44:57 -08:00
Krrish Dholakia
d2d57ecf1c test(test_parallel_request_limiter.py): fix test 2024-02-03 21:31:29 -08:00
Krrish Dholakia
3a19c8b600 test(test_completion.py): fix test 2024-02-03 21:30:45 -08:00
Krish Dholakia
28df60b609
Merge pull request #1809 from BerriAI/litellm_embedding_caching_updates
Support caching individual items in embedding list (Async embedding only)
2024-02-03 21:04:23 -08:00
ishaan-jaff
c353161456 (fix) test_parallel limiter fix 2024-02-03 21:03:15 -08:00
Krrish Dholakia
3e35041758 test(test_parallel_request_limiter.py): fix test to handle minute changes 2024-02-03 20:39:31 -08:00
ishaan-jaff
1155025e6a (ci/cd) run again 2024-02-03 20:36:35 -08:00
Krrish Dholakia
b47b2837eb test(test_parallel_request_limiter.py): fix test 2024-02-03 20:34:05 -08:00
ishaan-jaff
774cbbde52 (test) tgai is unstable 2024-02-03 20:00:40 -08:00
Krrish Dholakia
efb6123d28 fix(utils.py): support get_secret("TOGETHER_AI_TOKEN") 2024-02-03 19:35:09 -08:00
Krrish Dholakia
c49c88c8e5 fix(utils.py): route together ai calls to openai client
together ai is now openai-compatible

n
2024-02-03 19:22:48 -08:00
Krish Dholakia
9ab59045a3
Merge branch 'main' into litellm_embedding_caching_updates 2024-02-03 18:08:47 -08:00
Krish Dholakia
1d241b4001
Merge pull request #1802 from BerriAI/litellm_vertex_ai_high_traffic_fix
fix(vertex_ai.py): treat vertex ai high-traffic error as a rate limit error - allows user-controlled backoff logic to work here
2024-02-03 15:37:05 -08:00
Krish Dholakia
6408af11b6
Merge pull request #1799 from BerriAI/litellm_bedrock_stable_diffusion_support
feat(bedrock.py): add stable diffusion image generation support
2024-02-03 12:59:00 -08:00
Krrish Dholakia
e617ef981d fix(vertex_ai.py): treat vertex ai high-traffic error as a rate limit error - allows backoff logic to work here 2024-02-03 12:58:16 -08:00
Krish Dholakia
1c05d479e2
Merge pull request #1797 from BerriAI/litellm_vertex_ai_async_embedding_support
fix(vertex_ai.py): add async embedding support for vertex ai
2024-02-03 12:11:21 -08:00
Krrish Dholakia
36416360c4 feat(bedrock.py): add stable diffusion image generation support 2024-02-03 12:08:38 -08:00
Ishaan Jaff
aadc352cf0
Merge pull request #1795 from BerriAI/litellm_allow_setting_user_roles
[Feat] Allow setting user roles for UserTable
2024-02-03 11:13:09 -08:00
ishaan-jaff
8455aa4bd4 (test) setting user_role 2024-02-03 11:06:51 -08:00
Krrish Dholakia
0ffdf57dec fix(vertex_ai.py): add async embedding support for vertex ai 2024-02-03 10:35:17 -08:00
Krish Dholakia
5bf51a6058
Merge pull request #1794 from BerriAI/litellm_vertex_ai_gecko_support
Litellm vertex ai gecko support
2024-02-03 10:17:37 -08:00
Krrish Dholakia
1ba6882f76 fix(test_embedding.py): fix test 2024-02-03 09:49:23 -08:00
Krrish Dholakia
d9ba8668f4 feat(vertex_ai.py): vertex ai gecko text embedding support 2024-02-03 09:48:29 -08:00
Ishaan Jaff
6950f99804
Merge pull request #1788 from BerriAI/litellm_fix_delete_key
[Fix] /key/delete + add delete cache keys
2024-02-02 20:22:26 -08:00
ishaan-jaff
ae95e4c3f9 (test) user_api_key_auth_cache 2024-02-02 19:57:55 -08:00
Krish Dholakia
38a299a1bc
Merge pull request #1786 from BerriAI/litellm_cost_tracking_caching_fixes
Litellm cost tracking caching fixes (should be 0.0)
2024-02-02 19:23:07 -08:00
Krrish Dholakia
3aab719521 test(test_completion.py): skip flaky test 2024-02-02 19:03:47 -08:00
Krrish Dholakia
1f0598a277 fix(proxy_server.py): load default team config straight from config file 2024-02-02 19:03:47 -08:00
Krrish Dholakia
142e7cf1ce test(test_image_generation.py): ignore content violation errors for image gen test 2024-02-02 19:03:47 -08:00
Krrish Dholakia
c0699b08e5 test(test_caching.py): fix test to check id 2024-02-02 19:03:47 -08:00
Krrish Dholakia
c997d4b0ce fix(test_key_generate_prisma.py): add longer delay to allow logs to update 2024-02-02 19:03:47 -08:00
Krrish Dholakia
ddf3e51545 fix(main.py): for health checks, don't use cached responses 2024-02-02 19:03:47 -08:00