ishaan-jaff
f4fe2575cc
(fix) use client for text_completion()
2023-12-27 15:20:26 +05:30
Krrish Dholakia
c9fdbaf898
fix(azure.py,-openai.py): correctly raise errors if streaming calls fail
2023-12-27 15:08:37 +05:30
Krrish Dholakia
c88a8d71f0
fix: fix linting issues
2023-12-27 12:21:31 +05:30
Krish Dholakia
5c3a61d62f
Merge pull request #1248 from danikhan632/main
...
updated oobabooga to new api and support for embeddings
2023-12-27 11:33:56 +05:30
Ishaan Jaff
22d0c21829
Merge pull request #1249 from evantancy/main
...
fix: helicone logging
2023-12-27 11:24:19 +05:30
evantancy
668c786099
fix: helicone logging
2023-12-27 12:16:29 +08:00
dan
c4dfd9be7c
updated oobabooga to new api and support for embeddings
2023-12-26 19:45:28 -05:00
ishaan-jaff
751d57379d
(fix) support ollama_chat for acompletion
2023-12-26 20:01:51 +05:30
Krrish Dholakia
f0b6b9dce2
fix(main.py): support ttl being set for completion, embedding, image generation calls
2023-12-26 17:22:40 +05:30
ishaan-jaff
a463625452
(chore) completion - move functions lower
2023-12-26 14:35:59 +05:30
ishaan-jaff
7b097305c1
(feat) support logprobs, top_logprobs openai
2023-12-26 14:00:42 +05:30
ishaan-jaff
0b0d22d58c
(feat) add logprobs, top_logprobs to litellm.completion
2023-12-26 13:39:48 +05:30
ishaan-jaff
8c35aebdf8
(feat) ollama chat
2023-12-25 23:04:17 +05:30
ishaan-jaff
da4ec6c8b6
(feat) add ollama_chat v0
2023-12-25 14:27:10 +05:30
Krrish Dholakia
4905929de3
refactor: add black formatting
2023-12-25 14:11:20 +05:30
Krrish Dholakia
343a06fd84
fix(proxy_server.py): raise streaming exceptions
2023-12-25 07:18:09 +05:30
Krrish Dholakia
1262d89ab3
feat(gemini.py): add support for completion calls for gemini-pro (google ai studio)
2023-12-24 09:42:58 +05:30
Krrish Dholakia
eaaad79823
feat(ollama.py): add support for async ollama embeddings
2023-12-23 18:01:25 +05:30
Krrish Dholakia
be68796eba
fix(router.py): add support for async image generation endpoints
2023-12-21 14:38:44 +05:30
Mateo Cámara
b72d372aa7
feat: added explicit args to acomplete
2023-12-20 19:49:12 +01:00
Krrish Dholakia
f355e03515
feat(main.py): add async image generation support
2023-12-20 16:58:40 +05:30
Krrish Dholakia
b3962e483f
feat(azure.py): add support for azure image generations endpoint
2023-12-20 16:37:21 +05:30
Krrish Dholakia
f0df28362a
feat(ollama.py): add support for ollama function calling
2023-12-20 14:59:55 +05:30
Krish Dholakia
408f232bd7
Merge branch 'main' into main
2023-12-18 17:54:34 -08:00
Krrish Dholakia
34509d8dda
fix(main.py): return async completion calls
2023-12-18 17:41:54 -08:00
ishaan-jaff
bd15c61a65
(feat) OR default transforms=[]
2023-12-18 10:59:09 +05:30
ishaan-jaff
3a97a2817f
(fix) default args batch completion
2023-12-18 10:05:44 +05:30
ishaan-jaff
1e57c0c152
(feat) completion set function, function_call default None
2023-12-18 09:54:43 +05:30
ishaan-jaff
6b272076d7
(feat) openrouter set transforms=[]
default
2023-12-18 09:16:33 +05:30
Joel Eriksson
a419d59542
Fix for issue that occured when proxying to ollama
...
In the text_completion() function, it previously threw an exception at:
raw_response = response._hidden_params.get("original_response", None)
Due to response being an coroutine object to an ollama_acompletion call,
so I added an asyncio.iscoroutine() check for the response and handle it
by calling response = asyncio.run(response)
I also had to fix atext_completion(), where init_response was an instance
of TextCompletionResponse.
Since this case was not handled by the if-elif that checks if init_response
is a coroutine, a dict or a ModelResponse instance, response was unbound
which threw an exception on the "return response" line.
Note that a regular pyright based linter detects that response is possibly
unbound, and that the same code pattern is used in multiple other places
in main.py.
I would suggest that you either change these cases:
init_response = await loop.run_in_executor(...
if isinstance(init_response, ...
response = init_response
elif asyncio.iscoroutine(init_response):
response = await init_response
To either just:
response = await loop.run_in_executor(
if asyncio.iscoroutine(response):
response = await response
Or at the very least, include an else statement and set response = init_response,
so that response is never unbound when the code proceeds.
2023-12-17 17:27:47 +02:00
Krrish Dholakia
a3c7a340a5
fix(ollama.py): fix sync ollama streaming
2023-12-16 21:23:21 -08:00
Krrish Dholakia
13d088b72e
feat(main.py): add support for image generation endpoint
2023-12-16 21:07:29 -08:00
Krrish Dholakia
5b4ca42de6
docs(routing.md): add docs on using caching groups across deployments
2023-12-15 21:51:59 -08:00
Krrish Dholakia
84ad9f441e
feat(router.py): support caching groups
2023-12-15 21:45:51 -08:00
ishaan-jaff
287633887e
(feat) add ollama/llava
2023-12-16 10:35:27 +05:30
Krrish Dholakia
add153d110
fix(huggingface_restapi.py): add support for additional hf embedding formats
2023-12-15 21:02:41 -08:00
Krrish Dholakia
cab870f73a
fix(ollama.py): fix ollama async streaming for /completions calls
2023-12-15 09:28:32 -08:00
Krrish Dholakia
e45b491985
fix(main.py): log user for embedding calls
2023-12-14 15:14:37 -08:00
Krrish Dholakia
1608dd7e0b
fix(main.py): support async streaming for text completions endpoint
2023-12-14 13:56:32 -08:00
ishaan-jaff
7945664e61
(feat) add mistral api
2023-12-14 18:17:48 +05:30
ishaan-jaff
9526ee16c4
(fix) aembedding - don't pop aembedding out
2023-12-14 17:13:35 +05:30
Ishaan Jaff
cbc8cb08c8
Merge pull request #1112 from Undertone0809/add-cs-return-type-for-completion
...
refactor: add CustomStreamWrapper return type for completion
2023-12-14 16:52:11 +05:30
Krrish Dholakia
7b8851cce5
fix(ollama.py): fix async completion calls for ollama
2023-12-13 13:10:25 -08:00
Krrish Dholakia
69c29f8f86
fix(vertex_ai.py): add support for real async streaming + completion calls
2023-12-13 11:53:55 -08:00
Krrish Dholakia
07015843ac
fix(vertex_ai.py): support optional params + enable async calls for gemini
2023-12-13 11:01:23 -08:00
Krrish Dholakia
ef7a6e3ae1
feat(vertex_ai.py): adds support for gemini-pro on vertex ai
2023-12-13 10:26:30 -08:00
ishaan-jaff
86e626edab
(feat) pass vertex_ai/ as custom_llm_provider
2023-12-13 19:02:24 +03:00
zeeland
79ea466cf5
refactor: add CustomStreamWrapper return type for completion
2023-12-13 22:57:19 +08:00
Krrish Dholakia
8b07a6c046
fix(main.py): pass user_id + encoding_format for logging + to openai/azure
2023-12-12 15:46:44 -08:00
Krrish Dholakia
632d6e0bff
fix(utils.py): add more logging
2023-12-12 15:46:12 -08:00