mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-08 04:54:38 +00:00
Some checks failed
SqlStore Integration Tests / test-postgres (3.12) (push) Failing after 1s
SqlStore Integration Tests / test-postgres (3.13) (push) Failing after 0s
Integration Auth Tests / test-matrix (oauth2_token) (push) Failing after 1s
Test External Providers Installed via Module / test-external-providers-from-module (venv) (push) Has been skipped
Python Package Build Test / build (3.12) (push) Failing after 1s
Python Package Build Test / build (3.13) (push) Failing after 0s
Integration Tests (Replay) / Integration Tests (, , , client=, ) (push) Failing after 3s
Vector IO Integration Tests / test-matrix (push) Failing after 4s
Unit Tests / unit-tests (3.12) (push) Failing after 3s
Unit Tests / unit-tests (3.13) (push) Failing after 3s
Test External API and Providers / test-external (venv) (push) Failing after 5s
API Conformance Tests / check-schema-compatibility (push) Successful in 9s
UI Tests / ui-tests (22) (push) Successful in 43s
Pre-commit / pre-commit (push) Successful in 2m0s
# What does this PR do? * Fixes numerous broken links in the new documentation ## Test Plan * Server builds
134 lines
3.6 KiB
Text
134 lines
3.6 KiB
Text
---
|
|
title: Available Distributions
|
|
description: List of available distributions for Llama Stack
|
|
sidebar_label: Available Distributions
|
|
sidebar_position: 2
|
|
---
|
|
|
|
# Available Distributions
|
|
|
|
Llama Stack provides several pre-configured distributions to help you get started quickly. Choose the distribution that best fits your hardware and use case.
|
|
|
|
## Quick Reference
|
|
|
|
| Distribution | Use Case | Hardware Requirements | Provider |
|
|
|--------------|----------|----------------------|----------|
|
|
| `distribution-starter` | General purpose, prototyping | Any (CPU/GPU) | Ollama, Remote APIs |
|
|
| `distribution-meta-reference-gpu` | High-performance inference | GPU required | Local GPU inference |
|
|
| Remote-hosted | Production, managed service | None | Partner providers |
|
|
| iOS/Android SDK | Mobile applications | Mobile device | On-device inference |
|
|
|
|
## Choose Your Distribution
|
|
|
|
### 🚀 Getting Started (Recommended for Beginners)
|
|
|
|
**Use `distribution-starter` if you want to:**
|
|
- Prototype quickly without GPU requirements
|
|
- Use remote inference providers (Fireworks, Together, vLLM etc.)
|
|
- Run locally with Ollama for development
|
|
|
|
```bash
|
|
docker pull llama-stack/distribution-starter
|
|
```
|
|
|
|
**Guides:** [Starter Distribution Guide](self_hosted_distro/starter)
|
|
|
|
### 🖥️ Self-Hosted with GPU
|
|
|
|
**Use `distribution-meta-reference-gpu` if you:**
|
|
- Have access to GPU hardware
|
|
- Want maximum performance and control
|
|
- Need to run inference locally
|
|
|
|
```bash
|
|
docker pull llama-stack/distribution-meta-reference-gpu
|
|
```
|
|
|
|
**Guides:** [Meta Reference GPU Guide](self_hosted_distro/meta-reference-gpu)
|
|
|
|
### 🖥️ Self-Hosted with NVIDA NeMo Microservices
|
|
|
|
**Use `nvidia` if you:**
|
|
- Want to use Llama Stack with NVIDIA NeMo Microservices
|
|
|
|
**Guides:** [NVIDIA Distribution Guide](self_hosted_distro/nvidia)
|
|
|
|
### ☁️ Managed Hosting
|
|
|
|
**Use remote-hosted endpoints if you:**
|
|
- Don't want to manage infrastructure
|
|
- Need production-ready reliability
|
|
- Prefer managed services
|
|
|
|
**Partners:** [Fireworks.ai](https://fireworks.ai) and [Together.xyz](https://together.xyz)
|
|
|
|
**Guides:** [Remote-Hosted Endpoints](./remote_hosted_distro/)
|
|
|
|
### 📱 Mobile Development
|
|
|
|
**Use mobile SDKs if you:**
|
|
- Are building iOS or Android applications
|
|
- Need on-device inference capabilities
|
|
- Want offline functionality
|
|
|
|
- [iOS SDK](ondevice_distro/ios_sdk)
|
|
- [Android SDK](ondevice_distro/android_sdk)
|
|
|
|
### 🔧 Custom Solutions
|
|
|
|
**Build your own distribution if:**
|
|
- None of the above fit your specific needs
|
|
- You need custom configurations
|
|
- You want to optimize for your specific use case
|
|
|
|
**Guides:** [Building Custom Distributions](./building_distro)
|
|
|
|
## Detailed Documentation
|
|
|
|
### Self-Hosted Distributions
|
|
|
|
```{toctree}
|
|
:maxdepth: 1
|
|
|
|
self_hosted_distro/starter
|
|
self_hosted_distro/meta-reference-gpu
|
|
```
|
|
|
|
### Remote-Hosted Solutions
|
|
|
|
```{toctree}
|
|
:maxdepth: 1
|
|
|
|
remote_hosted_distro/index
|
|
```
|
|
|
|
### Mobile SDKs
|
|
|
|
```{toctree}
|
|
:maxdepth: 1
|
|
|
|
ondevice_distro/ios_sdk
|
|
ondevice_distro/android_sdk
|
|
```
|
|
|
|
## Decision Flow
|
|
|
|
```mermaid
|
|
graph TD
|
|
A[What's your use case?] --> B{Need mobile app?}
|
|
B -->|Yes| C[Use Mobile SDKs]
|
|
B -->|No| D{Have GPU hardware?}
|
|
D -->|Yes| E[Use Meta Reference GPU]
|
|
D -->|No| F{Want managed hosting?}
|
|
F -->|Yes| G[Use Remote-Hosted]
|
|
F -->|No| H[Use Starter Distribution]
|
|
```
|
|
|
|
## Next Steps
|
|
|
|
1. **Choose your distribution** from the options above
|
|
2. **Follow the setup guide** for your selected distribution
|
|
3. **Configure your providers** with API keys or local models
|
|
4. **Start building** with Llama Stack!
|
|
|
|
For help choosing or troubleshooting, check our [Getting Started Guide](/docs/getting_started/quickstart) or [Community Support](https://github.com/llamastack/llama-stack/discussions).
|