Krrish Dholakia
|
3c6bc031de
|
build(dockerfile): remove --config proxy_server_config.yaml from docker run
prevents startup errors with dockerfile
|
2024-04-08 13:23:56 -07:00 |
|
Krrish Dholakia
|
75d2eb61b4
|
bump: version 1.34.34 → 1.34.35
|
2024-04-08 12:19:46 -07:00 |
|
Krrish Dholakia
|
2fc169e6a0
|
refactor(main.py): trigger new build
|
2024-04-08 12:19:11 -07:00 |
|
Krrish Dholakia
|
48bfc45cb0
|
fix(utils.py): fix reordering of items for cached embeddings
ensures cached embedding item is returned in correct order
|
2024-04-08 12:18:24 -07:00 |
|
Ishaan Jaff
|
28e4706bfd
|
test - re-order embedding responses
|
2024-04-08 12:02:40 -07:00 |
|
Krrish Dholakia
|
d099591a09
|
docs(sidebars.js): refactor ordering
|
2024-04-08 07:30:08 -07:00 |
|
Ishaan Jaff
|
b4b882c5d6
|
bump: version 1.34.33 → 1.34.34
|
2024-04-07 09:57:27 -07:00 |
|
Ishaan Jaff
|
a5aef6ec00
|
Merge pull request #2879 from BerriAI/litellm_async_anthropic_api
[Feat] Async Anthropic API 97.5% lower median latency
|
2024-04-07 09:56:52 -07:00 |
|
Ishaan Jaff
|
3b6b497672
|
Merge pull request #2882 from BerriAI/litellm_docs_fix
docs fix gpt-3.5-turbo-instruct-0914
|
2024-04-06 20:26:24 -07:00 |
|
Krrish Dholakia
|
c60140a09a
|
bump: version 1.34.32 → 1.34.33
|
2024-04-06 19:44:29 -07:00 |
|
Krrish Dholakia
|
8a8b207fc9
|
refactor(main.py): trigger new build
|
2024-04-06 19:42:54 -07:00 |
|
Krrish Dholakia
|
6773341a08
|
test(test_alerting.py): remove duplicate test
|
2024-04-06 19:33:01 -07:00 |
|
Ishaan Jaff
|
d51e853b60
|
undo adding next-asyncio
|
2024-04-06 19:28:51 -07:00 |
|
Ishaan Jaff
|
9be250c0f0
|
add exit and aenter
|
2024-04-06 19:27:26 -07:00 |
|
Ishaan Jaff
|
3833c4083c
|
(ci/cd) fix queueing github releases
|
2024-04-06 19:21:58 -07:00 |
|
Krish Dholakia
|
83f608dc5d
|
Merge pull request #2880 from BerriAI/litellm_api_base_alerting
feat(proxy/utils.py): return api base for request hanging alerts
|
2024-04-06 19:17:18 -07:00 |
|
Ishaan Jaff
|
a38d3b17c5
|
ci/cd run async handler
|
2024-04-06 19:16:27 -07:00 |
|
Krrish Dholakia
|
d65efcce99
|
refactor(main.py): trigger new build
|
2024-04-06 18:50:38 -07:00 |
|
Krrish Dholakia
|
6234982fd9
|
test(test_custom_callback_input.py): fix assert
|
2024-04-06 18:48:33 -07:00 |
|
Krrish Dholakia
|
5f910c1811
|
test(test_streaming.py): fix vertex project name
|
2024-04-06 18:45:17 -07:00 |
|
Krrish Dholakia
|
460546956d
|
fix(utils.py): fix import
|
2024-04-06 18:37:38 -07:00 |
|
Krrish Dholakia
|
a410981972
|
fix(utils.py): fix circular import
|
2024-04-06 18:29:51 -07:00 |
|
Ishaan Jaff
|
f08486448c
|
fix - test streaming
|
2024-04-06 18:28:07 -07:00 |
|
Ishaan Jaff
|
2622f0351b
|
(ci/cd) run again
|
2024-04-06 18:26:52 -07:00 |
|
Krrish Dholakia
|
6f94f3d127
|
fix(router.py): improve pre-call check -> get model group cache one-time
|
2024-04-06 18:24:51 -07:00 |
|
Ishaan Jaff
|
4512510d37
|
Revert "fix test streaming"
This reverts commit f10aba1007 .
|
2024-04-06 18:24:00 -07:00 |
|
Ishaan Jaff
|
f10aba1007
|
fix test streaming
|
2024-04-06 18:23:10 -07:00 |
|
Ishaan Jaff
|
187774ef5c
|
bump: version 1.34.31 → 1.34.32
|
2024-04-06 18:20:23 -07:00 |
|
Ishaan Jaff
|
fcf5aa278b
|
fix - use anthropic class for clients
|
2024-04-06 18:19:28 -07:00 |
|
Krrish Dholakia
|
7ae6432f94
|
fix(router.py): check usage based routing cache in pre-call check
allows pre-call rpm check to work across instances
|
2024-04-06 18:19:02 -07:00 |
|
Ishaan Jaff
|
9be6b7ec7c
|
ci/cd run again
|
2024-04-06 18:07:41 -07:00 |
|
Krrish Dholakia
|
a6f987e065
|
test(test_amazing_vertex_completion.py): fix project name
|
2024-04-06 17:59:49 -07:00 |
|
Ishaan Jaff
|
548b2b6861
|
test - async claude streaming
|
2024-04-06 17:55:26 -07:00 |
|
Ishaan Jaff
|
2cf41d3d9f
|
async ahtropic streaming
|
2024-04-06 17:54:19 -07:00 |
|
Krrish Dholakia
|
205ac1496a
|
fix(router.py): store in-memory deployment request count for 60s only
|
2024-04-06 17:53:39 -07:00 |
|
Ishaan Jaff
|
5c796b4365
|
async streaming anthropic
|
2024-04-06 17:53:06 -07:00 |
|
Ishaan Jaff
|
7849c29f70
|
async anthropic streaming
|
2024-04-06 17:36:56 -07:00 |
|
Ishaan Jaff
|
8e5e99533b
|
async streaming for anthropic
|
2024-04-06 17:34:23 -07:00 |
|
Krrish Dholakia
|
0d1cca9aa0
|
fix(router.py): make router async calls coroutine safe
uses pre-call checks to check if a call is below it's rpm limit, works even if multiple async calls are
made simultaneously
|
2024-04-06 17:31:26 -07:00 |
|
Krish Dholakia
|
a101591f74
|
Update README.md
|
2024-04-06 16:11:33 -07:00 |
|
Krrish Dholakia
|
ead59b7b38
|
fix(utils.py): add gemini api base support to 'get_api_base'
|
2024-04-06 16:08:15 -07:00 |
|
Krrish Dholakia
|
6110d32b1c
|
feat(proxy/utils.py): return api base for request hanging alerts
|
2024-04-06 15:58:53 -07:00 |
|
Ishaan Jaff
|
58c4b02447
|
feat - make anthropic async
|
2024-04-06 15:50:13 -07:00 |
|
Ishaan Jaff
|
f09e6ba98d
|
Merge pull request #2878 from BerriAI/litellm_azure_gpt_instruct
re-use Azure OpenAI client for azure text completions
|
2024-04-06 13:35:39 -07:00 |
|
Ishaan Jaff
|
a6bc673ffa
|
feat - re-use OpenAI client for azure text
|
2024-04-06 12:23:58 -07:00 |
|
Ishaan Jaff
|
a2c63075ef
|
Merge pull request #2877 from BerriAI/litellm_fix_text_completion
[Feat] Text-Completion-OpenAI - Re-use OpenAI Client
|
2024-04-06 12:15:52 -07:00 |
|
Krrish Dholakia
|
b49e47b634
|
ci(vertex_key.json): switch gcloud accounts
|
2024-04-06 12:14:46 -07:00 |
|
Ishaan Jaff
|
c2f978fd5a
|
(docs) use text completion with litellm proxy
|
2024-04-06 12:07:20 -07:00 |
|
Ishaan Jaff
|
3b622556be
|
test - setting org for openai text completion
|
2024-04-06 11:53:26 -07:00 |
|
Ishaan Jaff
|
a55f3cdace
|
test - router re-use openai client
|
2024-04-06 11:33:17 -07:00 |
|