From 2153957ab6813889ed045c8f725a3882d029617b Mon Sep 17 00:00:00 2001 From: Young Han Date: Mon, 14 Jul 2025 17:52:27 -0700 Subject: [PATCH] docs --- docs/source/distributions/self_hosted_distro/llamacpp.md | 5 +++-- llama_stack/templates/llamacpp/doc_template.md | 5 +++-- 2 files changed, 6 insertions(+), 4 deletions(-) diff --git a/docs/source/distributions/self_hosted_distro/llamacpp.md b/docs/source/distributions/self_hosted_distro/llamacpp.md index e610aa695..c559a70ba 100644 --- a/docs/source/distributions/self_hosted_distro/llamacpp.md +++ b/docs/source/distributions/self_hosted_distro/llamacpp.md @@ -1,8 +1,9 @@ # Llama Stack with llama.cpp -This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. \n -Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. \n +This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. + +Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. You can employ any .gguf models available on [Hugging Face](https://huggingface.co/models) with this template. diff --git a/llama_stack/templates/llamacpp/doc_template.md b/llama_stack/templates/llamacpp/doc_template.md index f1a48c592..3b8ba8788 100644 --- a/llama_stack/templates/llamacpp/doc_template.md +++ b/llama_stack/templates/llamacpp/doc_template.md @@ -1,7 +1,8 @@ # Llama Stack with llama.cpp -This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. \n -Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. \n +This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. + +Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. You can employ any .gguf models available on [Hugging Face](https://huggingface.co/models) with this template.