mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-16 16:52:35 +00:00
2.5 KiB
2.5 KiB
Self-Hosted Distributions
:maxdepth: 2
:hidden:
meta-reference-gpu
meta-reference-quantized-gpu
ollama
tgi
dell-tgi
together
fireworks
remote-vllm
bedrock
We offer deployable distributions where you can host your own Llama Stack server using local inference.
| Distribution | Llama Stack Docker | Start This Distribution |
|---|---|---|
| Meta Reference | llamastack/distribution-meta-reference-gpu | Guide |
| Meta Reference Quantized | llamastack/distribution-meta-reference-quantized-gpu | Guide |
| Ollama | llamastack/distribution-ollama | Guide |
| TGI | llamastack/distribution-tgi | Guide |
| Together | llamastack/distribution-together | Guide |
| Fireworks | llamastack/distribution-fireworks | Guide |
| Bedrock | llamastack/distribution-bedrock | Guide |