mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-24 12:50:01 +00:00
docs
This commit is contained in:
parent
723a870171
commit
2153957ab6
2 changed files with 6 additions and 4 deletions
|
|
@ -1,8 +1,9 @@
|
|||
<!-- This file was auto-generated by distro_codegen.py, please edit source -->
|
||||
# Llama Stack with llama.cpp
|
||||
|
||||
This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider. \n
|
||||
Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp. \n
|
||||
This template demonstrates how to utilize Llama Stack with [llama.cpp](https://github.com/ggerganov/llama.cpp) as the inference provider.
|
||||
|
||||
Previously, the use of quantized models with Llama Stack was restricted, but now it is fully supported through llama.cpp.
|
||||
You can employ any .gguf models available on [Hugging Face](https://huggingface.co/models) with this template.
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue