llama-stack-mirror/distributions
2024-10-24 12:08:56 -07:00
..
bedrock llama stack distributions / templates / docker refactor (#266) 2024-10-21 11:17:53 -07:00
databricks llama stack distributions / templates / docker refactor (#266) 2024-10-21 11:17:53 -07:00
fireworks add more distro templates (#279) 2024-10-21 18:15:08 -07:00
hf-endpoint llama stack distributions / templates / docker refactor (#266) 2024-10-21 11:17:53 -07:00
hf-serverless llama stack distributions / templates / docker refactor (#266) 2024-10-21 11:17:53 -07:00
meta-reference-gpu use pytorch/pytorch as base 2024-10-23 20:22:00 -07:00
meta-reference-quantized-gpu Small updates to quantization config 2024-10-24 12:08:56 -07:00
ollama add more distro templates (#279) 2024-10-21 18:15:08 -07:00
tgi add more distro templates (#279) 2024-10-21 18:15:08 -07:00
together add more distro templates (#279) 2024-10-21 18:15:08 -07:00
vllm llama stack distributions / templates / docker refactor (#266) 2024-10-21 11:17:53 -07:00
README.md Add a meta-reference-quantized-gpu distribution 2024-10-23 21:45:50 -07:00

Llama Stack Distribution

A Distribution is where APIs and Providers are assembled together to provide a consistent whole to the end application developer. You can mix-and-match providers -- some could be backed by local code and some could be remote. As a hobbyist, you can serve a small model locally, but can choose a cloud provider for a large model. Regardless, the higher level APIs your app needs to work with don't need to change at all. You can even imagine moving across the server / mobile-device boundary as well always using the same uniform set of APIs for developing Generative AI applications.

Quick Start Llama Stack Distributions Guide

Distribution Llama Stack Docker Start This Distribution Inference Agents Memory Safety Telemetry
Meta Reference llamastack/distribution-meta-reference-gpu Guide meta-reference meta-reference meta-reference; remote::pgvector; remote::chromadb meta-reference meta-reference
Meta Reference Quantized llamastack/distribution-meta-reference-quantized-gpu Guide meta-reference-quantized meta-reference meta-reference; remote::pgvector; remote::chromadb meta-reference meta-reference
Ollama llamastack/distribution-ollama Guide remote::ollama meta-reference remote::pgvector; remote::chromadb remote::ollama meta-reference
TGI llamastack/distribution-tgi Guide remote::tgi meta-reference meta-reference; remote::pgvector; remote::chromadb meta-reference meta-reference
Together llamastack/distribution-together Guide remote::together meta-reference remote::weaviate meta-reference meta-reference
Fireworks llamastack/distribution-fireworks Guide remote::fireworks meta-reference remote::weaviate meta-reference meta-reference