llama-stack/llama_stack/providers/impls/meta_reference/inference
Xi Yan 07f9bf723f
fix broken --list-templates with adding build.yaml files for packaging (#327)
* add build files to templates

* fix templates

* manifest

* symlink

* symlink

* precommit

* change everything to docker build.yaml

* remove image_type in templates

* fix build from templates CLI

* fix readmes
2024-10-25 12:51:22 -07:00
..
quantization fix broken --list-templates with adding build.yaml files for packaging (#327) 2024-10-25 12:51:22 -07:00
__init__.py Split off meta-reference-quantized provider 2024-10-10 16:03:19 -07:00
config.py Allow overridding checkpoint_dir via config 2024-10-18 14:28:06 -07:00
generation.py New quantized models (#301) 2024-10-24 08:38:56 -07:00
inference.py Add support for Structured Output / Guided decoding (#281) 2024-10-22 12:53:34 -07:00
model_parallel.py Make all methods async def again; add completion() for meta-reference (#270) 2024-10-18 20:50:59 -07:00
parallel_utils.py Make all methods async def again; add completion() for meta-reference (#270) 2024-10-18 20:50:59 -07:00