mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 18:22:41 +00:00
1.3 KiB
1.3 KiB
Self-Hosted Distributions
:maxdepth: 1
:hidden:
ollama
tgi
remote-vllm
meta-reference-gpu
meta-reference-quantized-gpu
together
fireworks
bedrock
We offer deployable distributions where you can host your own Llama Stack server using local inference.
| Distribution | Llama Stack Docker | Start This Distribution |
|---|---|---|
| Ollama | {dockerhub}distribution-ollama |
Guide |
| TGI | {dockerhub}distribution-tgi |
Guide |
| vLLM | {dockerhub}distribution-remote-vllm |
Guide |
| Meta Reference | {dockerhub}distribution-meta-reference-gpu |
Guide |
| Meta Reference Quantized | {dockerhub}distribution-meta-reference-quantized-gpu |
Guide |
| Together | {dockerhub}distribution-together |
Guide |
| Fireworks | {dockerhub}distribution-fireworks |
Guide |
| Bedrock | {dockerhub}distribution-bedrock |
Guide |