llama-stack-mirror/distributions
2024-10-18 18:53:49 -07:00
..
bedrock move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
databricks move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
fireworks move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
hf-endpoint move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
hf-serverless move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
meta-reference-gpu rename 2024-10-18 17:28:26 -07:00
ollama rename 2024-10-18 17:28:26 -07:00
tgi rename 2024-10-18 17:28:26 -07:00
together move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
vllm move distribution/templates to distributions/ 2024-10-18 17:21:50 -07:00
README.md readme 2024-10-18 18:53:49 -07:00

Llama Stack Distribution

A Distribution is where APIs and Providers are assembled together to provide a consistent whole to the end application developer. You can mix-and-match providers -- some could be backed by local code and some could be remote. As a hobbyist, you can serve a small model locally, but can choose a cloud provider for a large model. Regardless, the higher level APIs your app needs to work with don't need to change at all. You can even imagine moving across the server / mobile-device boundary as well always using the same uniform set of APIs for developing Generative AI applications.

Supported Llama Stack Distributions

Distribution Llama Stack Docker Start This Distribution Inference Agents Memory Safety Telemetry
Meta Reference llamastack/distribution-meta-reference-gpu Guide ✔️ ✔️ ✔️ ✔️ ✔️
Ollama llamastack/distribution-ollama Guide ✔️ ✔️ ✔️ ✔️ ✔️
TGI llamastack/distribution-tgi Guide ✔️ ✔️ ✔️ ✔️ ✔️