forked from phoenix-oss/llama-stack-mirror
2.5 KiB
2.5 KiB
Self-Hosted Distributions
:maxdepth: 2
:hidden:
meta-reference-gpu
meta-reference-quantized-gpu
ollama
tgi
dell-tgi
together
fireworks
remote-vllm
bedrock
We offer deployable distributions where you can host your own Llama Stack server using local inference.
Distribution | Llama Stack Docker | Start This Distribution |
---|---|---|
Meta Reference | llamastack/distribution-meta-reference-gpu | Guide |
Meta Reference Quantized | llamastack/distribution-meta-reference-quantized-gpu | Guide |
Ollama | llamastack/distribution-ollama | Guide |
TGI | llamastack/distribution-tgi | Guide |
Together | llamastack/distribution-together | Guide |
Fireworks | llamastack/distribution-fireworks | Guide |
Bedrock | llamastack/distribution-bedrock | Guide |