Commit graph

423 commits

Author SHA1 Message Date
ishaan-jaff
c9be1cfcb1 (fix) ollama_chat - support function calling + fix for comp 2023-12-26 20:07:55 +05:30
ishaan-jaff
4233e42f5d (feat) ollama_chat acompletion without streaming 2023-12-26 20:01:51 +05:30
ishaan-jaff
dbf46823f8 (feat) ollama_chat add async stream 2023-12-25 23:45:27 +05:30
ishaan-jaff
b985d996b2 (feat) ollama_chat - add streaming support 2023-12-25 23:38:01 +05:30
ishaan-jaff
043d874ffe (feat) ollama/chat 2023-12-25 23:04:17 +05:30
ishaan-jaff
1742bd8716 (feat) ollama use /api/chat 2023-12-25 14:29:10 +05:30
ishaan-jaff
edf2b60765 (feat) add ollama_chat v0 2023-12-25 14:27:10 +05:30
Krrish Dholakia
79978c44ba refactor: add black formatting 2023-12-25 14:11:20 +05:30
Krrish Dholakia
70f4dabff6 feat(gemini.py): add support for completion calls for gemini-pro (google ai studio) 2023-12-24 09:42:58 +05:30
Krrish Dholakia
b7a7c3a4e5 feat(ollama.py): add support for async ollama embeddings 2023-12-23 18:01:25 +05:30
Krish Dholakia
4ec97e0c97 Merge pull request #1203 from Manouchehri/bedrock-cloudflare-ai-gateway-1
Add aws_bedrock_runtime_endpoint support
2023-12-23 11:44:04 +05:30
Krish Dholakia
36c1089029 Merge pull request #1213 from neubig/vertex_chat_generate_content
Make vertex ai work with generate_content
2023-12-23 11:40:43 +05:30
Krrish Dholakia
a65dfdde94 test(test_completion.py-+-test_streaming.py): add ollama endpoint to ci/cd pipeline 2023-12-22 12:21:33 +05:30
Krrish Dholakia
ae288c97fb fix(ollama.py): use litellm.request timeout for async call timeout 2023-12-22 11:22:24 +05:30
Graham Neubig
23b7a8a5ac Update the request_str 2023-12-21 09:58:06 -05:00
Krrish Dholakia
c084f04a35 fix(router.py): add support for async image generation endpoints 2023-12-21 14:38:44 +05:30
Krrish Dholakia
ee2a467708 fix(azure.py): correctly raise async exceptions 2023-12-21 12:23:07 +05:30
David Manouchehri
98bb605701 Add aws_bedrock_runtime_endpoint support. 2023-12-20 19:31:43 -05:00
Graham Neubig
b630528ef3 Make vertex_chat work with generate_content 2023-12-20 15:32:44 -05:00
Graham Neubig
c6b803ffaa Add a default for safety settings in vertex AI 2023-12-20 13:12:50 -05:00
Krrish Dholakia
b873833340 fix(router.py): only do sync image gen fallbacks for now
The customhttptransport we use for dall-e-2 only works for sync httpx calls, not async. Will need to spend some time writing the async version

n
2023-12-20 19:10:59 +05:30
Krrish Dholakia
a8f997eceb feat(main.py): add async image generation support 2023-12-20 16:58:40 +05:30
Krrish Dholakia
23d0278739 feat(azure.py): add support for azure image generations endpoint 2023-12-20 16:37:21 +05:30
Krrish Dholakia
636ac9b605 feat(ollama.py): add support for ollama function calling 2023-12-20 14:59:55 +05:30
Graham Neubig
efb4af930b Add partial support of vertexai safety settings 2023-12-19 22:26:55 -05:00
ishaan-jaff
3c37e0d58b (fix) proxy + ollama - raise exception correctly 2023-12-19 18:48:34 +05:30
Krish Dholakia
7e3f9d344c Merge branch 'main' into main 2023-12-18 17:54:34 -08:00
Krrish Dholakia
e03713ef74 fix(main.py): return async completion calls 2023-12-18 17:41:54 -08:00
ishaan-jaff
759a4702d7 (feat) openrouter set transforms=[] default 2023-12-18 09:16:33 +05:30
ishaan-jaff
d3c1c4bf28 (feat) set default openrouter configs 2023-12-18 08:55:51 +05:30
Joel Eriksson
afcc83bb15 Fix bug when iterating over lines in ollama response
async for line in resp.content.iter_any() will return
incomplete lines when the lines are long, and that
results in an exception being thrown by json.loads()
when it tries to parse the incomplete JSON

The default behavior of the stream reader for aiohttp
response objects is to iterate over lines, so just
removing .iter_any() fixes the bug
2023-12-17 20:23:26 +02:00
Krrish Dholakia
5f4310f592 fix(ollama.py): fix sync ollama streaming 2023-12-16 21:23:21 -08:00
Krrish Dholakia
51cb16a015 feat(main.py): add support for image generation endpoint 2023-12-16 21:07:29 -08:00
Krrish Dholakia
bfafabea10 fix(azure.py): fix azure streaming logging 2023-12-16 18:06:08 -08:00
Krrish Dholakia
c87cdbd156 build(Dockerfile): fixing build requirements 2023-12-16 17:52:30 -08:00
Krrish Dholakia
87df233a19 fix(health.md): add background health check details to docs 2023-12-16 10:31:59 -08:00
ishaan-jaff
a97aa37d84 (fix) vertexai - gemini 2023-12-16 22:15:41 +05:30
ishaan-jaff
5a5ab37560 (feat) add async, async+stream for gemini 2023-12-16 18:58:12 +05:30
ishaan-jaff
ea91cd7b83 (fix) use litellm.vertex_vision_models 2023-12-16 18:39:40 +05:30
ishaan-jaff
03b55820a1 (feat) add vertex ai gemini-pro-vision 2023-12-16 18:31:03 +05:30
ishaan-jaff
7a1e8b4bf3 (feat) show POST request for HF embeddings 2023-12-16 13:09:49 +05:30
ishaan-jaff
90bb6d4caf (feat) add ollama/llava 2023-12-16 10:35:27 +05:30
Krrish Dholakia
726289cb49 fix(huggingface_restapi.py): add support for additional hf embedding formats 2023-12-15 21:02:41 -08:00
Krrish Dholakia
1da7d35218 feat(proxy_server.py): enable infinite retries on rate limited requests 2023-12-15 20:03:41 -08:00
Krrish Dholakia
6081602bf8 fix(router.py): support openai-compatible endpoints 2023-12-15 14:47:54 -08:00
Krrish Dholakia
e3d0f0df50 fix(together_ai.py): return empty tgai responses 2023-12-15 10:46:35 -08:00
Krrish Dholakia
dad0b064e9 fix(together_ai.py): additional logging for together ai encoding prompt 2023-12-15 10:39:23 -08:00
Krrish Dholakia
3d6ade8f26 fix(ollama.py): fix ollama async streaming for /completions calls 2023-12-15 09:28:32 -08:00
ishaan-jaff
fb2d3c9aa6 (feat) - acompletion, correct exception mapping 2023-12-15 08:28:12 +05:30
Krrish Dholakia
a207f8ea63 bump: version 1.14.4 → 1.14.5.dev1 2023-12-14 15:23:52 -08:00