mogith-pn
|
318b4813f2
|
Clarifai-LiteLLM integration (#1)
* intg v1 clarifai-litellm
* Added more community models and testcase
* Clarifai-updated markdown docs
|
2024-04-30 22:38:33 +05:30 |
|
Krrish Dholakia
|
060ac995d6
|
fix(vertex_ai.py): accept credentials as a json string
|
2024-04-16 17:34:25 -07:00 |
|
Krrish Dholakia
|
c66b59a71e
|
refactor(main.py): trigger new build
|
2024-04-15 18:36:51 -07:00 |
|
Krrish Dholakia
|
1ec7118e1f
|
fix(vertex_ai_anthropic.py): set vertex_credentials for vertex ai anthropic calls
allows setting vertex credentials as a json string for vertex ai anthropic calls
|
2024-04-15 14:16:28 -07:00 |
|
Krrish Dholakia
|
50081479f9
|
fix(main.py): accept vertex service account credentials as json string
allows us to dynamically set vertex ai credentials
|
2024-04-15 13:28:59 -07:00 |
|
Krrish Dholakia
|
26286a54b8
|
fix(anthropic_text.py): add support for async text completion calls
|
2024-04-15 08:15:00 -07:00 |
|
Krrish Dholakia
|
c08d6a961a
|
refactor(main.py): trigger new build
|
2024-04-13 19:35:39 -07:00 |
|
Krrish Dholakia
|
74aa230eac
|
fix(main.py): automatically infer mode for text completion models
|
2024-04-12 14:16:21 -07:00 |
|
Krrish Dholakia
|
623613203a
|
refactor(main.py): trigger new build
contains fixes for async batch get
|
2024-04-10 21:45:06 -07:00 |
|
Ishaan Jaff
|
bc50b0a4a1
|
Merge pull request #2923 from BerriAI/litellm_return_better_error_from_health
fix - return stack trace on failing /health checks - first 1000 chars
|
2024-04-10 17:48:13 -07:00 |
|
Krrish Dholakia
|
a943f21f75
|
refactor(main.py): trigger new build
|
2024-04-09 21:15:33 -07:00 |
|
Krrish Dholakia
|
855e7ed9d2
|
fix(main.py): handle translating text completion openai to chat completion for async requests
also adds testing for this, to prevent future regressions
|
2024-04-09 16:47:49 -07:00 |
|
Ishaan Jaff
|
c9108d43e0
|
fix - return stack trace on failing /health checks
|
2024-04-09 15:12:09 -07:00 |
|
Krrish Dholakia
|
f2e34f456d
|
fix(main.py): trigger new build
|
2024-04-08 14:33:38 -07:00 |
|
Krrish Dholakia
|
2fc169e6a0
|
refactor(main.py): trigger new build
|
2024-04-08 12:19:11 -07:00 |
|
Ishaan Jaff
|
a5aef6ec00
|
Merge pull request #2879 from BerriAI/litellm_async_anthropic_api
[Feat] Async Anthropic API 97.5% lower median latency
|
2024-04-07 09:56:52 -07:00 |
|
Krrish Dholakia
|
8a8b207fc9
|
refactor(main.py): trigger new build
|
2024-04-06 19:42:54 -07:00 |
|
Krrish Dholakia
|
d65efcce99
|
refactor(main.py): trigger new build
|
2024-04-06 18:50:38 -07:00 |
|
Ishaan Jaff
|
fcf5aa278b
|
fix - use anthropic class for clients
|
2024-04-06 18:19:28 -07:00 |
|
Ishaan Jaff
|
8e5e99533b
|
async streaming for anthropic
|
2024-04-06 17:34:23 -07:00 |
|
Ishaan Jaff
|
58c4b02447
|
feat - make anthropic async
|
2024-04-06 15:50:13 -07:00 |
|
Ishaan Jaff
|
a2c63075ef
|
Merge pull request #2877 from BerriAI/litellm_fix_text_completion
[Feat] Text-Completion-OpenAI - Re-use OpenAI Client
|
2024-04-06 12:15:52 -07:00 |
|
Ishaan Jaff
|
ad7302cdc8
|
feat - re-use openai client for text completion
|
2024-04-06 11:25:33 -07:00 |
|
Krrish Dholakia
|
ef9e29d575
|
refactor(main.py): trigger new build
|
2024-04-06 09:06:53 -07:00 |
|
Krrish Dholakia
|
645fd62c15
|
refactor(main.py): trigger new build
|
2024-04-05 13:42:56 -07:00 |
|
Krrish Dholakia
|
695eb129ad
|
fix(router.py): fix client init for streaming timeouts
|
2024-04-05 12:30:15 -07:00 |
|
Krish Dholakia
|
eb34306099
|
Merge pull request #2665 from BerriAI/litellm_claude_vertex_ai
[WIP] feat(vertex_ai_anthropic.py): Add support for claude 3 on vertex ai
|
2024-04-05 07:06:04 -07:00 |
|
Krrish Dholakia
|
1441fccf9a
|
fix(main.py): trigger new build
|
2024-04-04 13:36:21 -07:00 |
|
Krrish Dholakia
|
f88545261c
|
build(config.yml): add missing dep
|
2024-04-04 10:59:26 -07:00 |
|
Krrish Dholakia
|
98fa3ff1c5
|
refactor(main.py): trigger new build
|
2024-04-04 10:54:13 -07:00 |
|
Krrish Dholakia
|
62b8749fe8
|
build: trigger new build
|
2024-04-04 10:23:13 -07:00 |
|
Krrish Dholakia
|
7f3a8d2a5e
|
build: trigger new build
|
2024-04-04 10:20:25 -07:00 |
|
Krrish Dholakia
|
a26732e710
|
refactor(main.py): trigger new build
|
2024-04-03 08:01:26 -07:00 |
|
Krrish Dholakia
|
88e8f14b69
|
fix(main.py): support async calls from azure_text
|
2024-04-03 07:59:32 -07:00 |
|
Krrish Dholakia
|
1d341970ba
|
feat(vertex_ai_anthropic.py): add claude 3 on vertex ai support - working .completions call
.completions() call works
|
2024-04-02 22:07:39 -07:00 |
|
Krrish Dholakia
|
b07788d2a5
|
fix(openai.py): return logprobs for text completion calls
|
2024-04-02 14:05:56 -07:00 |
|
Krrish Dholakia
|
72abe200a9
|
fix(main.py): fix elif block
|
2024-04-02 09:47:49 -07:00 |
|
Krish Dholakia
|
d95a1f3a28
|
Merge pull request #2790 from phact/patch-2
Fix max_tokens type in main.py
|
2024-04-02 09:02:34 -07:00 |
|
Krrish Dholakia
|
2fc7aede12
|
refactor(main.py): trigger new build
|
2024-04-02 08:51:18 -07:00 |
|
Krrish Dholakia
|
0d949d71ab
|
fix(main.py): support text completion input being a list of strings
addresses - https://github.com/BerriAI/litellm/issues/2792, https://github.com/BerriAI/litellm/issues/2777
|
2024-04-02 08:50:16 -07:00 |
|
Sebastián Estévez
|
d3a86e7b7f
|
Fix max_tokens type in main.py
|
2024-04-02 00:28:08 -04:00 |
|
Krrish Dholakia
|
ceabf726b0
|
fix(main.py): support max retries for transcription calls
|
2024-04-01 18:37:53 -07:00 |
|
Krrish Dholakia
|
ca54b62656
|
refactor(main.py): trigger new build
|
2024-04-01 18:03:46 -07:00 |
|
Krrish Dholakia
|
f5d920e314
|
refactor(main.py): trigger new build
|
2024-03-30 21:41:14 -07:00 |
|
Krrish Dholakia
|
c0204310ee
|
fix(main.py): fix translation to text_completions format for async text completion calls
|
2024-03-30 09:02:51 -07:00 |
|
Krrish Dholakia
|
63271846c2
|
refactor(main.py): trigger new build
|
2024-03-29 09:24:47 -07:00 |
|
Krrish Dholakia
|
d547944556
|
fix(sagemaker.py): support 'model_id' param for sagemaker
allow passing inference component param to sagemaker in the same format as we handle this for bedrock
|
2024-03-29 08:43:17 -07:00 |
|
Krish Dholakia
|
28905c85b6
|
Merge pull request #2720 from onukura/ollama-batch-embedding
Batch embedding for Ollama
|
2024-03-28 14:58:55 -07:00 |
|
Krrish Dholakia
|
664663f301
|
refactor(main.py): trigger new build
|
2024-03-28 14:52:47 -07:00 |
|
onukura
|
f86472518d
|
Add a feature to ollama aembedding to accept batch input
|
2024-03-27 21:39:19 +00:00 |
|