ishaan-jaff
c9be1cfcb1
(fix) ollama_chat - support function calling + fix for comp
2023-12-26 20:07:55 +05:30
ishaan-jaff
4233e42f5d
(feat) ollama_chat acompletion without streaming
2023-12-26 20:01:51 +05:30
ishaan-jaff
dbf46823f8
(feat) ollama_chat add async stream
2023-12-25 23:45:27 +05:30
ishaan-jaff
b985d996b2
(feat) ollama_chat - add streaming support
2023-12-25 23:38:01 +05:30
ishaan-jaff
043d874ffe
(feat) ollama/chat
2023-12-25 23:04:17 +05:30
ishaan-jaff
1742bd8716
(feat) ollama use /api/chat
2023-12-25 14:29:10 +05:30
ishaan-jaff
edf2b60765
(feat) add ollama_chat v0
2023-12-25 14:27:10 +05:30
Krrish Dholakia
79978c44ba
refactor: add black formatting
2023-12-25 14:11:20 +05:30
Krrish Dholakia
70f4dabff6
feat(gemini.py): add support for completion calls for gemini-pro (google ai studio)
2023-12-24 09:42:58 +05:30
Krrish Dholakia
b7a7c3a4e5
feat(ollama.py): add support for async ollama embeddings
2023-12-23 18:01:25 +05:30
Krish Dholakia
4ec97e0c97
Merge pull request #1203 from Manouchehri/bedrock-cloudflare-ai-gateway-1
...
Add aws_bedrock_runtime_endpoint support
2023-12-23 11:44:04 +05:30
Krish Dholakia
36c1089029
Merge pull request #1213 from neubig/vertex_chat_generate_content
...
Make vertex ai work with generate_content
2023-12-23 11:40:43 +05:30
Krrish Dholakia
a65dfdde94
test(test_completion.py-+-test_streaming.py): add ollama endpoint to ci/cd pipeline
2023-12-22 12:21:33 +05:30
Krrish Dholakia
ae288c97fb
fix(ollama.py): use litellm.request timeout for async call timeout
2023-12-22 11:22:24 +05:30
Graham Neubig
23b7a8a5ac
Update the request_str
2023-12-21 09:58:06 -05:00
Krrish Dholakia
c084f04a35
fix(router.py): add support for async image generation endpoints
2023-12-21 14:38:44 +05:30
Krrish Dholakia
ee2a467708
fix(azure.py): correctly raise async exceptions
2023-12-21 12:23:07 +05:30
David Manouchehri
98bb605701
Add aws_bedrock_runtime_endpoint support.
2023-12-20 19:31:43 -05:00
Graham Neubig
b630528ef3
Make vertex_chat work with generate_content
2023-12-20 15:32:44 -05:00
Graham Neubig
c6b803ffaa
Add a default for safety settings in vertex AI
2023-12-20 13:12:50 -05:00
Krrish Dholakia
b873833340
fix(router.py): only do sync image gen fallbacks for now
...
The customhttptransport we use for dall-e-2 only works for sync httpx calls, not async. Will need to spend some time writing the async version
n
2023-12-20 19:10:59 +05:30
Krrish Dholakia
a8f997eceb
feat(main.py): add async image generation support
2023-12-20 16:58:40 +05:30
Krrish Dholakia
23d0278739
feat(azure.py): add support for azure image generations endpoint
2023-12-20 16:37:21 +05:30
Krrish Dholakia
636ac9b605
feat(ollama.py): add support for ollama function calling
2023-12-20 14:59:55 +05:30
Graham Neubig
efb4af930b
Add partial support of vertexai safety settings
2023-12-19 22:26:55 -05:00
ishaan-jaff
3c37e0d58b
(fix) proxy + ollama - raise exception correctly
2023-12-19 18:48:34 +05:30
Krish Dholakia
7e3f9d344c
Merge branch 'main' into main
2023-12-18 17:54:34 -08:00
Krrish Dholakia
e03713ef74
fix(main.py): return async completion calls
2023-12-18 17:41:54 -08:00
ishaan-jaff
759a4702d7
(feat) openrouter set transforms=[]
default
2023-12-18 09:16:33 +05:30
ishaan-jaff
d3c1c4bf28
(feat) set default openrouter configs
2023-12-18 08:55:51 +05:30
Joel Eriksson
afcc83bb15
Fix bug when iterating over lines in ollama response
...
async for line in resp.content.iter_any() will return
incomplete lines when the lines are long, and that
results in an exception being thrown by json.loads()
when it tries to parse the incomplete JSON
The default behavior of the stream reader for aiohttp
response objects is to iterate over lines, so just
removing .iter_any() fixes the bug
2023-12-17 20:23:26 +02:00
Krrish Dholakia
5f4310f592
fix(ollama.py): fix sync ollama streaming
2023-12-16 21:23:21 -08:00
Krrish Dholakia
51cb16a015
feat(main.py): add support for image generation endpoint
2023-12-16 21:07:29 -08:00
Krrish Dholakia
bfafabea10
fix(azure.py): fix azure streaming logging
2023-12-16 18:06:08 -08:00
Krrish Dholakia
c87cdbd156
build(Dockerfile): fixing build requirements
2023-12-16 17:52:30 -08:00
Krrish Dholakia
87df233a19
fix(health.md): add background health check details to docs
2023-12-16 10:31:59 -08:00
ishaan-jaff
a97aa37d84
(fix) vertexai - gemini
2023-12-16 22:15:41 +05:30
ishaan-jaff
5a5ab37560
(feat) add async, async+stream for gemini
2023-12-16 18:58:12 +05:30
ishaan-jaff
ea91cd7b83
(fix) use litellm.vertex_vision_models
2023-12-16 18:39:40 +05:30
ishaan-jaff
03b55820a1
(feat) add vertex ai gemini-pro-vision
2023-12-16 18:31:03 +05:30
ishaan-jaff
7a1e8b4bf3
(feat) show POST request for HF embeddings
2023-12-16 13:09:49 +05:30
ishaan-jaff
90bb6d4caf
(feat) add ollama/llava
2023-12-16 10:35:27 +05:30
Krrish Dholakia
726289cb49
fix(huggingface_restapi.py): add support for additional hf embedding formats
2023-12-15 21:02:41 -08:00
Krrish Dholakia
1da7d35218
feat(proxy_server.py): enable infinite retries on rate limited requests
2023-12-15 20:03:41 -08:00
Krrish Dholakia
6081602bf8
fix(router.py): support openai-compatible endpoints
2023-12-15 14:47:54 -08:00
Krrish Dholakia
e3d0f0df50
fix(together_ai.py): return empty tgai responses
2023-12-15 10:46:35 -08:00
Krrish Dholakia
dad0b064e9
fix(together_ai.py): additional logging for together ai encoding prompt
2023-12-15 10:39:23 -08:00
Krrish Dholakia
3d6ade8f26
fix(ollama.py): fix ollama async streaming for /completions calls
2023-12-15 09:28:32 -08:00
ishaan-jaff
fb2d3c9aa6
(feat) - acompletion, correct exception mapping
2023-12-15 08:28:12 +05:30
Krrish Dholakia
a207f8ea63
bump: version 1.14.4 → 1.14.5.dev1
2023-12-14 15:23:52 -08:00