Commit graph

441 commits

Author SHA1 Message Date
ishaan-jaff
6672591198 (fix) init_bedrock_client 2024-01-01 22:48:56 +05:30
Krrish Dholakia
7be5f74b70 fix(aimage_generation): fix response type 2023-12-30 12:53:24 +05:30
Krrish Dholakia
4d239f1e65 fix(openai.py): fix async image gen call 2023-12-30 12:44:54 +05:30
Krrish Dholakia
b69ffb3738 fix: support dynamic timeouts for openai and azure 2023-12-30 12:14:02 +05:30
Krrish Dholakia
7d55a563ee fix(main.py): don't set timeout as an optional api param 2023-12-30 11:47:07 +05:30
ishaan-jaff
224d38ba48 (fix) vertex ai - use usage from response 2023-12-29 16:30:25 +05:30
ishaan-jaff
c69f4f17a5 (feat) cloudflare - add optional params 2023-12-29 11:50:09 +05:30
ishaan-jaff
b990fc8324 (feat) cloudflare ai workers - add completion support 2023-12-29 11:34:58 +05:30
ishaan-jaff
796e735881 (feat) v0 adding cloudflare 2023-12-29 09:32:29 +05:30
ishaan-jaff
362bed6ca3 (fix) together_ai cost tracking 2023-12-28 22:11:08 +05:30
Krrish Dholakia
5a48dac83f fix(vertex_ai.py): support function calling for gemini 2023-12-28 19:07:04 +05:30
Krrish Dholakia
8188475c16 feat(admin_ui.py): support creating keys on admin ui 2023-12-28 16:59:11 +05:30
Krrish Dholakia
507b6bf96e fix(utils.py): use local tiktoken copy 2023-12-28 11:22:33 +05:30
Krrish Dholakia
2285282ef8 feat(health_check.py): more detailed health check calls 2023-12-28 09:12:57 +05:30
Krrish Dholakia
db6ef70a68 fix(azure.py): raise streaming exceptions 2023-12-27 15:43:13 +05:30
Krrish Dholakia
fd5e6efb1d fix(azure.py,-openai.py): correctly raise errors if streaming calls fail 2023-12-27 15:08:37 +05:30
Krrish Dholakia
2269f01c17 fix: fix linting issues 2023-12-27 12:21:31 +05:30
dan
c7be18cf46 updated oobabooga to new api and support for embeddings 2023-12-26 19:45:28 -05:00
ishaan-jaff
c9be1cfcb1 (fix) ollama_chat - support function calling + fix for comp 2023-12-26 20:07:55 +05:30
ishaan-jaff
4233e42f5d (feat) ollama_chat acompletion without streaming 2023-12-26 20:01:51 +05:30
ishaan-jaff
dbf46823f8 (feat) ollama_chat add async stream 2023-12-25 23:45:27 +05:30
ishaan-jaff
b985d996b2 (feat) ollama_chat - add streaming support 2023-12-25 23:38:01 +05:30
ishaan-jaff
043d874ffe (feat) ollama/chat 2023-12-25 23:04:17 +05:30
ishaan-jaff
1742bd8716 (feat) ollama use /api/chat 2023-12-25 14:29:10 +05:30
ishaan-jaff
edf2b60765 (feat) add ollama_chat v0 2023-12-25 14:27:10 +05:30
Krrish Dholakia
79978c44ba refactor: add black formatting 2023-12-25 14:11:20 +05:30
Krrish Dholakia
70f4dabff6 feat(gemini.py): add support for completion calls for gemini-pro (google ai studio) 2023-12-24 09:42:58 +05:30
Krrish Dholakia
b7a7c3a4e5 feat(ollama.py): add support for async ollama embeddings 2023-12-23 18:01:25 +05:30
Krish Dholakia
4ec97e0c97 Merge pull request #1203 from Manouchehri/bedrock-cloudflare-ai-gateway-1
Add aws_bedrock_runtime_endpoint support
2023-12-23 11:44:04 +05:30
Krish Dholakia
36c1089029 Merge pull request #1213 from neubig/vertex_chat_generate_content
Make vertex ai work with generate_content
2023-12-23 11:40:43 +05:30
Krrish Dholakia
a65dfdde94 test(test_completion.py-+-test_streaming.py): add ollama endpoint to ci/cd pipeline 2023-12-22 12:21:33 +05:30
Krrish Dholakia
ae288c97fb fix(ollama.py): use litellm.request timeout for async call timeout 2023-12-22 11:22:24 +05:30
Graham Neubig
23b7a8a5ac Update the request_str 2023-12-21 09:58:06 -05:00
Krrish Dholakia
c084f04a35 fix(router.py): add support for async image generation endpoints 2023-12-21 14:38:44 +05:30
Krrish Dholakia
ee2a467708 fix(azure.py): correctly raise async exceptions 2023-12-21 12:23:07 +05:30
David Manouchehri
98bb605701 Add aws_bedrock_runtime_endpoint support. 2023-12-20 19:31:43 -05:00
Graham Neubig
b630528ef3 Make vertex_chat work with generate_content 2023-12-20 15:32:44 -05:00
Graham Neubig
c6b803ffaa Add a default for safety settings in vertex AI 2023-12-20 13:12:50 -05:00
Krrish Dholakia
b873833340 fix(router.py): only do sync image gen fallbacks for now
The customhttptransport we use for dall-e-2 only works for sync httpx calls, not async. Will need to spend some time writing the async version

n
2023-12-20 19:10:59 +05:30
Krrish Dholakia
a8f997eceb feat(main.py): add async image generation support 2023-12-20 16:58:40 +05:30
Krrish Dholakia
23d0278739 feat(azure.py): add support for azure image generations endpoint 2023-12-20 16:37:21 +05:30
Krrish Dholakia
636ac9b605 feat(ollama.py): add support for ollama function calling 2023-12-20 14:59:55 +05:30
Graham Neubig
efb4af930b Add partial support of vertexai safety settings 2023-12-19 22:26:55 -05:00
ishaan-jaff
3c37e0d58b (fix) proxy + ollama - raise exception correctly 2023-12-19 18:48:34 +05:30
Krish Dholakia
7e3f9d344c Merge branch 'main' into main 2023-12-18 17:54:34 -08:00
Krrish Dholakia
e03713ef74 fix(main.py): return async completion calls 2023-12-18 17:41:54 -08:00
ishaan-jaff
759a4702d7 (feat) openrouter set transforms=[] default 2023-12-18 09:16:33 +05:30
ishaan-jaff
d3c1c4bf28 (feat) set default openrouter configs 2023-12-18 08:55:51 +05:30
Joel Eriksson
afcc83bb15 Fix bug when iterating over lines in ollama response
async for line in resp.content.iter_any() will return
incomplete lines when the lines are long, and that
results in an exception being thrown by json.loads()
when it tries to parse the incomplete JSON

The default behavior of the stream reader for aiohttp
response objects is to iterate over lines, so just
removing .iter_any() fixes the bug
2023-12-17 20:23:26 +02:00
Krrish Dholakia
5f4310f592 fix(ollama.py): fix sync ollama streaming 2023-12-16 21:23:21 -08:00