Commit graph

9746 commits

Author SHA1 Message Date
Krrish Dholakia
3c6bc031de build(dockerfile): remove --config proxy_server_config.yaml from docker run
prevents startup errors with dockerfile
2024-04-08 13:23:56 -07:00
Krrish Dholakia
75d2eb61b4 bump: version 1.34.34 → 1.34.35 2024-04-08 12:19:46 -07:00
Krrish Dholakia
2fc169e6a0 refactor(main.py): trigger new build 2024-04-08 12:19:11 -07:00
Krrish Dholakia
48bfc45cb0 fix(utils.py): fix reordering of items for cached embeddings
ensures cached embedding item is returned in correct order
2024-04-08 12:18:24 -07:00
Ishaan Jaff
28e4706bfd test - re-order embedding responses 2024-04-08 12:02:40 -07:00
Krrish Dholakia
d099591a09 docs(sidebars.js): refactor ordering 2024-04-08 07:30:08 -07:00
Ishaan Jaff
b4b882c5d6 bump: version 1.34.33 → 1.34.34 2024-04-07 09:57:27 -07:00
Ishaan Jaff
a5aef6ec00
Merge pull request #2879 from BerriAI/litellm_async_anthropic_api
[Feat] Async Anthropic API 97.5% lower median latency
2024-04-07 09:56:52 -07:00
Ishaan Jaff
3b6b497672
Merge pull request #2882 from BerriAI/litellm_docs_fix
docs fix gpt-3.5-turbo-instruct-0914
2024-04-06 20:26:24 -07:00
Krrish Dholakia
c60140a09a bump: version 1.34.32 → 1.34.33 2024-04-06 19:44:29 -07:00
Krrish Dholakia
8a8b207fc9 refactor(main.py): trigger new build 2024-04-06 19:42:54 -07:00
Krrish Dholakia
6773341a08 test(test_alerting.py): remove duplicate test 2024-04-06 19:33:01 -07:00
Ishaan Jaff
d51e853b60 undo adding next-asyncio 2024-04-06 19:28:51 -07:00
Ishaan Jaff
9be250c0f0 add exit and aenter 2024-04-06 19:27:26 -07:00
Ishaan Jaff
3833c4083c (ci/cd) fix queueing github releases 2024-04-06 19:21:58 -07:00
Krish Dholakia
83f608dc5d
Merge pull request #2880 from BerriAI/litellm_api_base_alerting
feat(proxy/utils.py): return api base for request hanging alerts
2024-04-06 19:17:18 -07:00
Ishaan Jaff
a38d3b17c5 ci/cd run async handler 2024-04-06 19:16:27 -07:00
Krrish Dholakia
d65efcce99 refactor(main.py): trigger new build 2024-04-06 18:50:38 -07:00
Krrish Dholakia
6234982fd9 test(test_custom_callback_input.py): fix assert 2024-04-06 18:48:33 -07:00
Krrish Dholakia
5f910c1811 test(test_streaming.py): fix vertex project name 2024-04-06 18:45:17 -07:00
Krrish Dholakia
460546956d fix(utils.py): fix import 2024-04-06 18:37:38 -07:00
Krrish Dholakia
a410981972 fix(utils.py): fix circular import 2024-04-06 18:29:51 -07:00
Ishaan Jaff
f08486448c fix - test streaming 2024-04-06 18:28:07 -07:00
Ishaan Jaff
2622f0351b (ci/cd) run again 2024-04-06 18:26:52 -07:00
Krrish Dholakia
6f94f3d127 fix(router.py): improve pre-call check -> get model group cache one-time 2024-04-06 18:24:51 -07:00
Ishaan Jaff
4512510d37 Revert "fix test streaming"
This reverts commit f10aba1007.
2024-04-06 18:24:00 -07:00
Ishaan Jaff
f10aba1007 fix test streaming 2024-04-06 18:23:10 -07:00
Ishaan Jaff
187774ef5c bump: version 1.34.31 → 1.34.32 2024-04-06 18:20:23 -07:00
Ishaan Jaff
fcf5aa278b fix - use anthropic class for clients 2024-04-06 18:19:28 -07:00
Krrish Dholakia
7ae6432f94 fix(router.py): check usage based routing cache in pre-call check
allows pre-call rpm check to work across instances
2024-04-06 18:19:02 -07:00
Ishaan Jaff
9be6b7ec7c ci/cd run again 2024-04-06 18:07:41 -07:00
Krrish Dholakia
a6f987e065 test(test_amazing_vertex_completion.py): fix project name 2024-04-06 17:59:49 -07:00
Ishaan Jaff
548b2b6861 test - async claude streaming 2024-04-06 17:55:26 -07:00
Ishaan Jaff
2cf41d3d9f async ahtropic streaming 2024-04-06 17:54:19 -07:00
Krrish Dholakia
205ac1496a fix(router.py): store in-memory deployment request count for 60s only 2024-04-06 17:53:39 -07:00
Ishaan Jaff
5c796b4365 async streaming anthropic 2024-04-06 17:53:06 -07:00
Ishaan Jaff
7849c29f70 async anthropic streaming 2024-04-06 17:36:56 -07:00
Ishaan Jaff
8e5e99533b async streaming for anthropic 2024-04-06 17:34:23 -07:00
Krrish Dholakia
0d1cca9aa0 fix(router.py): make router async calls coroutine safe
uses pre-call checks to check if a call is below it's rpm limit, works even if multiple async calls are
made simultaneously
2024-04-06 17:31:26 -07:00
Krish Dholakia
a101591f74
Update README.md 2024-04-06 16:11:33 -07:00
Krrish Dholakia
ead59b7b38 fix(utils.py): add gemini api base support to 'get_api_base' 2024-04-06 16:08:15 -07:00
Krrish Dholakia
6110d32b1c feat(proxy/utils.py): return api base for request hanging alerts 2024-04-06 15:58:53 -07:00
Ishaan Jaff
58c4b02447 feat - make anthropic async 2024-04-06 15:50:13 -07:00
Ishaan Jaff
f09e6ba98d
Merge pull request #2878 from BerriAI/litellm_azure_gpt_instruct
re-use Azure OpenAI client for azure text completions
2024-04-06 13:35:39 -07:00
Ishaan Jaff
a6bc673ffa feat - re-use OpenAI client for azure text 2024-04-06 12:23:58 -07:00
Ishaan Jaff
a2c63075ef
Merge pull request #2877 from BerriAI/litellm_fix_text_completion
[Feat] Text-Completion-OpenAI - Re-use OpenAI Client
2024-04-06 12:15:52 -07:00
Krrish Dholakia
b49e47b634 ci(vertex_key.json): switch gcloud accounts 2024-04-06 12:14:46 -07:00
Ishaan Jaff
c2f978fd5a (docs) use text completion with litellm proxy 2024-04-06 12:07:20 -07:00
Ishaan Jaff
3b622556be test - setting org for openai text completion 2024-04-06 11:53:26 -07:00
Ishaan Jaff
a55f3cdace test - router re-use openai client 2024-04-06 11:33:17 -07:00