ishaan-jaff
6672591198
(fix) init_bedrock_client
2024-01-01 22:48:56 +05:30
Krrish Dholakia
7be5f74b70
fix(aimage_generation): fix response type
2023-12-30 12:53:24 +05:30
Krrish Dholakia
4d239f1e65
fix(openai.py): fix async image gen call
2023-12-30 12:44:54 +05:30
Krrish Dholakia
b69ffb3738
fix: support dynamic timeouts for openai and azure
2023-12-30 12:14:02 +05:30
Krrish Dholakia
7d55a563ee
fix(main.py): don't set timeout as an optional api param
2023-12-30 11:47:07 +05:30
ishaan-jaff
224d38ba48
(fix) vertex ai - use usage from response
2023-12-29 16:30:25 +05:30
ishaan-jaff
c69f4f17a5
(feat) cloudflare - add optional params
2023-12-29 11:50:09 +05:30
ishaan-jaff
b990fc8324
(feat) cloudflare ai workers - add completion support
2023-12-29 11:34:58 +05:30
ishaan-jaff
796e735881
(feat) v0 adding cloudflare
2023-12-29 09:32:29 +05:30
ishaan-jaff
362bed6ca3
(fix) together_ai cost tracking
2023-12-28 22:11:08 +05:30
Krrish Dholakia
5a48dac83f
fix(vertex_ai.py): support function calling for gemini
2023-12-28 19:07:04 +05:30
Krrish Dholakia
8188475c16
feat(admin_ui.py): support creating keys on admin ui
2023-12-28 16:59:11 +05:30
Krrish Dholakia
507b6bf96e
fix(utils.py): use local tiktoken copy
2023-12-28 11:22:33 +05:30
Krrish Dholakia
2285282ef8
feat(health_check.py): more detailed health check calls
2023-12-28 09:12:57 +05:30
Krrish Dholakia
db6ef70a68
fix(azure.py): raise streaming exceptions
2023-12-27 15:43:13 +05:30
Krrish Dholakia
fd5e6efb1d
fix(azure.py,-openai.py): correctly raise errors if streaming calls fail
2023-12-27 15:08:37 +05:30
Krrish Dholakia
2269f01c17
fix: fix linting issues
2023-12-27 12:21:31 +05:30
dan
c7be18cf46
updated oobabooga to new api and support for embeddings
2023-12-26 19:45:28 -05:00
ishaan-jaff
c9be1cfcb1
(fix) ollama_chat - support function calling + fix for comp
2023-12-26 20:07:55 +05:30
ishaan-jaff
4233e42f5d
(feat) ollama_chat acompletion without streaming
2023-12-26 20:01:51 +05:30
ishaan-jaff
dbf46823f8
(feat) ollama_chat add async stream
2023-12-25 23:45:27 +05:30
ishaan-jaff
b985d996b2
(feat) ollama_chat - add streaming support
2023-12-25 23:38:01 +05:30
ishaan-jaff
043d874ffe
(feat) ollama/chat
2023-12-25 23:04:17 +05:30
ishaan-jaff
1742bd8716
(feat) ollama use /api/chat
2023-12-25 14:29:10 +05:30
ishaan-jaff
edf2b60765
(feat) add ollama_chat v0
2023-12-25 14:27:10 +05:30
Krrish Dholakia
79978c44ba
refactor: add black formatting
2023-12-25 14:11:20 +05:30
Krrish Dholakia
70f4dabff6
feat(gemini.py): add support for completion calls for gemini-pro (google ai studio)
2023-12-24 09:42:58 +05:30
Krrish Dholakia
b7a7c3a4e5
feat(ollama.py): add support for async ollama embeddings
2023-12-23 18:01:25 +05:30
Krish Dholakia
4ec97e0c97
Merge pull request #1203 from Manouchehri/bedrock-cloudflare-ai-gateway-1
...
Add aws_bedrock_runtime_endpoint support
2023-12-23 11:44:04 +05:30
Krish Dholakia
36c1089029
Merge pull request #1213 from neubig/vertex_chat_generate_content
...
Make vertex ai work with generate_content
2023-12-23 11:40:43 +05:30
Krrish Dholakia
a65dfdde94
test(test_completion.py-+-test_streaming.py): add ollama endpoint to ci/cd pipeline
2023-12-22 12:21:33 +05:30
Krrish Dholakia
ae288c97fb
fix(ollama.py): use litellm.request timeout for async call timeout
2023-12-22 11:22:24 +05:30
Graham Neubig
23b7a8a5ac
Update the request_str
2023-12-21 09:58:06 -05:00
Krrish Dholakia
c084f04a35
fix(router.py): add support for async image generation endpoints
2023-12-21 14:38:44 +05:30
Krrish Dholakia
ee2a467708
fix(azure.py): correctly raise async exceptions
2023-12-21 12:23:07 +05:30
David Manouchehri
98bb605701
Add aws_bedrock_runtime_endpoint support.
2023-12-20 19:31:43 -05:00
Graham Neubig
b630528ef3
Make vertex_chat work with generate_content
2023-12-20 15:32:44 -05:00
Graham Neubig
c6b803ffaa
Add a default for safety settings in vertex AI
2023-12-20 13:12:50 -05:00
Krrish Dholakia
b873833340
fix(router.py): only do sync image gen fallbacks for now
...
The customhttptransport we use for dall-e-2 only works for sync httpx calls, not async. Will need to spend some time writing the async version
n
2023-12-20 19:10:59 +05:30
Krrish Dholakia
a8f997eceb
feat(main.py): add async image generation support
2023-12-20 16:58:40 +05:30
Krrish Dholakia
23d0278739
feat(azure.py): add support for azure image generations endpoint
2023-12-20 16:37:21 +05:30
Krrish Dholakia
636ac9b605
feat(ollama.py): add support for ollama function calling
2023-12-20 14:59:55 +05:30
Graham Neubig
efb4af930b
Add partial support of vertexai safety settings
2023-12-19 22:26:55 -05:00
ishaan-jaff
3c37e0d58b
(fix) proxy + ollama - raise exception correctly
2023-12-19 18:48:34 +05:30
Krish Dholakia
7e3f9d344c
Merge branch 'main' into main
2023-12-18 17:54:34 -08:00
Krrish Dholakia
e03713ef74
fix(main.py): return async completion calls
2023-12-18 17:41:54 -08:00
ishaan-jaff
759a4702d7
(feat) openrouter set transforms=[]
default
2023-12-18 09:16:33 +05:30
ishaan-jaff
d3c1c4bf28
(feat) set default openrouter configs
2023-12-18 08:55:51 +05:30
Joel Eriksson
afcc83bb15
Fix bug when iterating over lines in ollama response
...
async for line in resp.content.iter_any() will return
incomplete lines when the lines are long, and that
results in an exception being thrown by json.loads()
when it tries to parse the incomplete JSON
The default behavior of the stream reader for aiohttp
response objects is to iterate over lines, so just
removing .iter_any() fixes the bug
2023-12-17 20:23:26 +02:00
Krrish Dholakia
5f4310f592
fix(ollama.py): fix sync ollama streaming
2023-12-16 21:23:21 -08:00