diff --git a/docs/source/distributions/self_hosted_distro/llamacpp.md b/docs/source/distributions/self_hosted_distro/llamacpp.md index e610aa695..c559a70ba 100644 --- a/docs/source/distributions/self_hosted_distro/llamacpp.md +++ b/docs/source/distributions/self_hosted_distro/llamacpp.md @@ -1,8 +1,9 @@ # Llama Stack with llama.cpp -This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. \n -Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. \n +This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. + +Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. You can employ any .gguf models available on [Hugging Face](https://huggingface.co/models) with this template. diff --git a/llama_stack/templates/llamacpp/doc_template.md b/llama_stack/templates/llamacpp/doc_template.md index f1a48c592..3b8ba8788 100644 --- a/llama_stack/templates/llamacpp/doc_template.md +++ b/llama_stack/templates/llamacpp/doc_template.md @@ -1,7 +1,8 @@ # Llama Stack with llama.cpp -This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. \n -Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. \n +This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. + +Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. You can employ any .gguf models available on [Hugging Face](https://huggingface.co/models) with this template.