Commit graph

70 commits

Author SHA1 Message Date
Krrish Dholakia
311521e56e fix(ollama.py): correctly raise ollama streaming error
Fixes https://github.com/BerriAI/litellm/issues/4974
2024-07-30 15:01:26 -07:00
Titusz
fcef2c4580
Add missing num_gpu ollama configuration parameter 2024-07-18 17:51:56 +02:00
Krrish Dholakia
6e9f048618 fix: move to using pydantic obj for setting values 2024-07-11 13:18:36 -07:00
corrm
423a60c8bc chore: Improved OllamaConfig get_required_params and ollama_acompletion and ollama_async_streaming functions 2024-06-24 05:55:22 +03:00
Krish Dholakia
677e0255c8
Merge branch 'main' into litellm_cleanup_traceback 2024-06-06 16:32:08 -07:00
Krrish Dholakia
6cca5612d2 refactor: replace 'traceback.print_exc()' with logging library
allows error logs to be in json format for otel logging
2024-06-06 13:47:43 -07:00
sha-ahammed
faa4dfe03e feat: Add Ollama as a provider in the proxy UI 2024-06-05 16:48:38 +05:30
KX
d3921a3d28 fix: add missing seed parameter to ollama input
Current ollama interfacing does not allow for seed, which is supported in https://github.com/ollama/ollama/blob/main/docs/api.md#parameters and https://github.com/ollama/ollama/blob/main/docs/modelfile.md#valid-parameters-and-values

This resolves that by adding in handling of seed parameter.
2024-05-31 01:47:56 +08:00
frob
c44970c813
Merge branch 'BerriAI:main' into ollama-image-handling 2024-05-09 20:25:30 +02:00
Krrish Dholakia
6575143460 feat(proxy_server.py): return litellm version in response headers 2024-05-08 16:00:08 -07:00
frob
b93c00abec
Merge branch 'BerriAI:main' into ollama-image-handling 2024-05-09 00:14:29 +02:00
Ishaan Jaff
2725a55e7a
Merge pull request #3470 from mbektas/fix-ollama-embeddings
support sync ollama embeddings
2024-05-07 19:21:37 -07:00
frob
7a1a3f6411
Merge branch 'BerriAI:main' into ollama-image-handling 2024-05-06 18:06:45 +02:00
Mehmet Bektas
3acad270e5 support sync ollama embeddings 2024-05-05 19:44:25 -07:00
Jack Collins
bb6132eee1 Fix: get format from data not optional_params ollama non-stream completion 2024-05-05 18:59:26 -07:00
Jack Collins
81b1c46c6f Add missing import itertools.chain 2024-05-05 18:54:08 -07:00
Jack Collins
03b82b78c1 Fix: Set finish_reason to tool_calls for non-stream responses in ollama 2024-05-05 18:52:31 -07:00
Jack Collins
297543e3e5 Parse streamed function calls as single delta in ollama 2024-05-05 18:52:20 -07:00
frob
465f491e7f
Merge branch 'BerriAI:main' into ollama-image-handling 2024-05-01 22:29:37 +02:00
Krish Dholakia
0714eb3526
Merge branch 'main' into litellm_ollama_tool_call_reponse 2024-05-01 10:24:05 -07:00
frob
ae87cb3a31
Merge branch 'BerriAI:main' into ollama-image-handling 2024-04-21 01:49:10 +02:00
frob
3df7231fa5
Disable special tokens in ollama completion when counting tokens
Some(?) models (eg, codegemma) don't return a prompt_eval_count field, so ollama.py tries to compute the value based on encoding of the prompt.  Unfortunately FIM symbols used in the prompt (eg, "<|fim_prefix|>") cause the encoder to throw an exception, so we disable special processing.
2024-04-19 21:38:42 +02:00
frob
2492fade3a
Update comment. 2024-04-16 01:12:24 +02:00
frob
ea117fc859
Merge branch 'BerriAI:main' into ollama-image-handling 2024-04-13 21:42:58 +02:00
frob
82a4232dce
ollama also accepts PNG 2024-04-08 03:35:02 +02:00
frob
59ed4fb51e
Update ollama.py for image handling
ollama wants plain base64 jpeg images, and some clients send dataURI and/or webp.  Remove prefixes and convert all non-jpeg images to jpeg.
2024-04-08 03:28:24 +02:00
Gregory Nwosu
559a4cde23
created defaults for response["eval_count"]
there is no way in litellm to disable the cache in ollama that is removing the eval_count response keys from the json.
This PR allows the code to create sensible defaults for when the response is empty
see 
- https://github.com/ollama/ollama/issues/1573
- https://github.com/ollama/ollama/issues/2023
2024-04-08 02:03:54 +01:00
frob
d5c1ae1cb2
Update ollama.py for image handling
Some clients (eg librechat) send images in datauri format, not plain base64.  Strip off the prerix when passing images to ollama.
2024-04-07 13:05:39 +02:00
DaxServer
61b6f8be44 docs: Update references to Ollama repository url
Updated references to the Ollama repository URL from https://github.com/jmorganca/ollama to https://github.com/ollama/ollama.
2024-03-31 19:35:37 +02:00
Krrish Dholakia
48af367885 fix(ollama.py): fix type issue 2024-03-28 15:01:56 -07:00
onukura
f86472518d Add a feature to ollama aembedding to accept batch input 2024-03-27 21:39:19 +00:00
onukura
2df63cc621 Fix ollama embedding response 2024-03-25 16:26:49 +00:00
Lunik
cee20695eb
🐛 fix: Ollama vision models call arguments (like : llava)
Signed-off-by: Lunik <lunik@tiwabbit.fr>
2024-02-26 17:52:55 +01:00
Krrish Dholakia
d1db67890c fix(ollama.py): support format for ollama 2024-02-06 10:11:52 -08:00
Ishaan Jaff
14c9e239a1
Merge pull request #1750 from vanpelt/patch-2
Re-raise exception in async ollama streaming
2024-02-05 08:12:17 -08:00
Krrish Dholakia
312c7462c8 refactor(ollama.py): trigger rebuild 2024-02-03 20:23:43 -08:00
Krrish Dholakia
01cef1fe9e fix(ollama.py): fix api connection error
https://github.com/BerriAI/litellm/issues/1735
2024-02-03 20:22:33 -08:00
Chris Van Pelt
1568b162f5
Re-raise exception in async ollama streaming 2024-02-01 16:14:07 -08:00
Krrish Dholakia
d46df34ff5 fix(utils.py): fix streaming chunks to not return role, unless set 2024-02-01 09:55:56 -08:00
TheDiscoMole
ed07de2729 changing ollama response parsing to expected behaviour 2024-01-19 23:36:24 +01:00
ishaan-jaff
5f2cbfc711 (feat) litellm.completion - support ollama timeout 2024-01-09 10:34:41 +05:30
Krrish Dholakia
88d498a54a fix(ollama.py): use tiktoken as backup for prompt token counting 2024-01-09 09:47:18 +05:30
Krrish Dholakia
4905929de3 refactor: add black formatting 2023-12-25 14:11:20 +05:30
Krrish Dholakia
eaaad79823 feat(ollama.py): add support for async ollama embeddings 2023-12-23 18:01:25 +05:30
Krrish Dholakia
eb2d13e2fb test(test_completion.py-+-test_streaming.py): add ollama endpoint to ci/cd pipeline 2023-12-22 12:21:33 +05:30
Krrish Dholakia
57607f111a fix(ollama.py): use litellm.request timeout for async call timeout 2023-12-22 11:22:24 +05:30
Krrish Dholakia
f0df28362a feat(ollama.py): add support for ollama function calling 2023-12-20 14:59:55 +05:30
ishaan-jaff
9995229b97 (fix) proxy + ollama - raise exception correctly 2023-12-19 18:48:34 +05:30
Joel Eriksson
e214e6ab47 Fix bug when iterating over lines in ollama response
async for line in resp.content.iter_any() will return
incomplete lines when the lines are long, and that
results in an exception being thrown by json.loads()
when it tries to parse the incomplete JSON

The default behavior of the stream reader for aiohttp
response objects is to iterate over lines, so just
removing .iter_any() fixes the bug
2023-12-17 20:23:26 +02:00
Krrish Dholakia
a3c7a340a5 fix(ollama.py): fix sync ollama streaming 2023-12-16 21:23:21 -08:00