Commit graph

773 commits

Author SHA1 Message Date
Krrish Dholakia
74aa230eac fix(main.py): automatically infer mode for text completion models 2024-04-12 14:16:21 -07:00
Krrish Dholakia
623613203a refactor(main.py): trigger new build
contains fixes for async batch get
2024-04-10 21:45:06 -07:00
Ishaan Jaff
bc50b0a4a1
Merge pull request #2923 from BerriAI/litellm_return_better_error_from_health
fix - return stack trace on failing /health checks - first 1000 chars
2024-04-10 17:48:13 -07:00
Krrish Dholakia
a943f21f75 refactor(main.py): trigger new build 2024-04-09 21:15:33 -07:00
Krrish Dholakia
855e7ed9d2 fix(main.py): handle translating text completion openai to chat completion for async requests
also adds testing for this, to prevent future regressions
2024-04-09 16:47:49 -07:00
Ishaan Jaff
c9108d43e0 fix - return stack trace on failing /health checks 2024-04-09 15:12:09 -07:00
Krrish Dholakia
f2e34f456d fix(main.py): trigger new build 2024-04-08 14:33:38 -07:00
Krrish Dholakia
2fc169e6a0 refactor(main.py): trigger new build 2024-04-08 12:19:11 -07:00
Ishaan Jaff
a5aef6ec00
Merge pull request #2879 from BerriAI/litellm_async_anthropic_api
[Feat] Async Anthropic API 97.5% lower median latency
2024-04-07 09:56:52 -07:00
Krrish Dholakia
8a8b207fc9 refactor(main.py): trigger new build 2024-04-06 19:42:54 -07:00
Krrish Dholakia
d65efcce99 refactor(main.py): trigger new build 2024-04-06 18:50:38 -07:00
Ishaan Jaff
fcf5aa278b fix - use anthropic class for clients 2024-04-06 18:19:28 -07:00
Ishaan Jaff
8e5e99533b async streaming for anthropic 2024-04-06 17:34:23 -07:00
Ishaan Jaff
58c4b02447 feat - make anthropic async 2024-04-06 15:50:13 -07:00
Ishaan Jaff
a2c63075ef
Merge pull request #2877 from BerriAI/litellm_fix_text_completion
[Feat] Text-Completion-OpenAI - Re-use OpenAI Client
2024-04-06 12:15:52 -07:00
Ishaan Jaff
ad7302cdc8 feat - re-use openai client for text completion 2024-04-06 11:25:33 -07:00
Krrish Dholakia
ef9e29d575 refactor(main.py): trigger new build 2024-04-06 09:06:53 -07:00
Krrish Dholakia
645fd62c15 refactor(main.py): trigger new build 2024-04-05 13:42:56 -07:00
Krrish Dholakia
695eb129ad fix(router.py): fix client init for streaming timeouts 2024-04-05 12:30:15 -07:00
Krish Dholakia
eb34306099
Merge pull request #2665 from BerriAI/litellm_claude_vertex_ai
[WIP] feat(vertex_ai_anthropic.py): Add support for claude 3 on vertex ai
2024-04-05 07:06:04 -07:00
Krrish Dholakia
1441fccf9a fix(main.py): trigger new build 2024-04-04 13:36:21 -07:00
Krrish Dholakia
f88545261c build(config.yml): add missing dep 2024-04-04 10:59:26 -07:00
Krrish Dholakia
98fa3ff1c5 refactor(main.py): trigger new build 2024-04-04 10:54:13 -07:00
Krrish Dholakia
62b8749fe8 build: trigger new build 2024-04-04 10:23:13 -07:00
Krrish Dholakia
7f3a8d2a5e build: trigger new build 2024-04-04 10:20:25 -07:00
Krrish Dholakia
a26732e710 refactor(main.py): trigger new build 2024-04-03 08:01:26 -07:00
Krrish Dholakia
88e8f14b69 fix(main.py): support async calls from azure_text 2024-04-03 07:59:32 -07:00
Krrish Dholakia
1d341970ba feat(vertex_ai_anthropic.py): add claude 3 on vertex ai support - working .completions call
.completions() call works
2024-04-02 22:07:39 -07:00
Krrish Dholakia
b07788d2a5 fix(openai.py): return logprobs for text completion calls 2024-04-02 14:05:56 -07:00
Krrish Dholakia
72abe200a9 fix(main.py): fix elif block 2024-04-02 09:47:49 -07:00
Krish Dholakia
d95a1f3a28
Merge pull request #2790 from phact/patch-2
Fix max_tokens type in main.py
2024-04-02 09:02:34 -07:00
Krrish Dholakia
2fc7aede12 refactor(main.py): trigger new build 2024-04-02 08:51:18 -07:00
Krrish Dholakia
0d949d71ab fix(main.py): support text completion input being a list of strings
addresses - https://github.com/BerriAI/litellm/issues/2792, https://github.com/BerriAI/litellm/issues/2777
2024-04-02 08:50:16 -07:00
Sebastián Estévez
d3a86e7b7f
Fix max_tokens type in main.py 2024-04-02 00:28:08 -04:00
Krrish Dholakia
ceabf726b0 fix(main.py): support max retries for transcription calls 2024-04-01 18:37:53 -07:00
Krrish Dholakia
ca54b62656 refactor(main.py): trigger new build 2024-04-01 18:03:46 -07:00
Krrish Dholakia
f5d920e314 refactor(main.py): trigger new build 2024-03-30 21:41:14 -07:00
Krrish Dholakia
c0204310ee fix(main.py): fix translation to text_completions format for async text completion calls 2024-03-30 09:02:51 -07:00
Krrish Dholakia
63271846c2 refactor(main.py): trigger new build 2024-03-29 09:24:47 -07:00
Krrish Dholakia
d547944556 fix(sagemaker.py): support 'model_id' param for sagemaker
allow passing inference component param to sagemaker in the same format as we handle this for bedrock
2024-03-29 08:43:17 -07:00
Krish Dholakia
28905c85b6
Merge pull request #2720 from onukura/ollama-batch-embedding
Batch embedding for Ollama
2024-03-28 14:58:55 -07:00
Krrish Dholakia
664663f301 refactor(main.py): trigger new build 2024-03-28 14:52:47 -07:00
onukura
f86472518d Add a feature to ollama aembedding to accept batch input 2024-03-27 21:39:19 +00:00
Krrish Dholakia
9375b131ee refactor(main.py): trigger new build 2024-03-26 21:18:51 -07:00
Krish Dholakia
7eb2c7942c
Merge pull request #2675 from onukura/ollama-embedding
Fix Ollama embedding
2024-03-26 16:08:28 -07:00
onukura
ef69eefcdb Fix ollama api_base to enable remote url 2024-03-25 16:26:40 +00:00
Krrish Dholakia
f98aead602 feat(main.py): support router.chat.completions.create
allows using router with instructor

https://github.com/BerriAI/litellm/issues/2673
2024-03-25 08:26:28 -07:00
Krrish Dholakia
b7321ae4ee fix(router.py): fix pre call check logic 2024-03-23 18:56:08 -07:00
Krrish Dholakia
94f55aa6d9 fix(bedrock.py): support claude 3 function calling when stream=true
https://github.com/BerriAI/litellm/issues/2615
2024-03-21 18:39:03 -07:00
Krrish Dholakia
af27a61d76 refactor(main.py): trigger new build 2024-03-21 10:56:44 -07:00