llama-stack-mirror/llama_toolchain
Hardik Shah 156bfa0e15
Added Ollama as an inference impl (#20)
* fix non-streaming api in inference server

* unit test for inline inference

* Added non-streaming ollama inference impl

* add streaming support for ollama inference with tests

* addressing comments

---------

Co-authored-by: Hardik Shah <hjshah@fb.com>
2024-07-31 22:08:37 -07:00
..
cli update toolchain to work with updated imports from llama_models 2024-07-30 17:52:57 -07:00
common Initial commit 2024-07-23 08:32:33 -07:00
data Initial commit 2024-07-23 08:32:33 -07:00
dataset/api Initial commit 2024-07-23 08:32:33 -07:00
evaluations/api Initial commit 2024-07-23 08:32:33 -07:00
inference Added Ollama as an inference impl (#20) 2024-07-31 22:08:37 -07:00
memory Initial commit 2024-07-23 08:32:33 -07:00
models/api Initial commit 2024-07-23 08:32:33 -07:00
post_training/api Initial commit 2024-07-23 08:32:33 -07:00
reward_scoring/api Initial commit 2024-07-23 08:32:33 -07:00
safety Begin adding a /safety/run_shield API 2024-07-31 21:57:10 -07:00
synthetic_data_generation/api Initial commit 2024-07-23 08:32:33 -07:00
__init__.py Initial commit 2024-07-23 08:32:33 -07:00
utils.py Initial commit 2024-07-23 08:32:33 -07:00