llama-stack-mirror/docs
Matthew Farrellee 8cdfdbe884 feat: Add S3 Files Provider implementation
Implements a complete S3-based file storage provider for Llama Stack with:

Core Implementation:
- S3FilesImpl class with full OpenAI Files API compatibility
- Support for file upload, download, listing, deletion operations
- Sqlite-based metadata storage for fast queries and API compliance
- Configurable S3 endpoints (AWS, MinIO, LocalStack support)

Key Features:
- Automatic S3 bucket creation and management
- Metadata persistence
- Proper error handling for S3 connectivity and permissions

Dependencies:
- Adds boto3 for AWS S3 integration
- Adds moto[s3] for testing infrastructure

Testing:

 Unit: `./scripts/unit-tests.sh tests/unit/files tests/unit/providers/files`

 Integration:

  Start MinIO: `podman run --rm -it -p 9000:9000 minio/minio server /data`

  Start stack w/ S3 provider: `S3_ENDPOINT_URL=http://localhost:9000 AWS_ACCESS_KEY_ID=minioadmin AWS_SECRET_ACCESS_KEY=minioadmin S3_BUCKET_NAME=llama-stack-files uv run llama stack build --image-type venv --providers files=remote::s3 --run`

  Run integration tests: `./scripts/integration-tests.sh --stack-config http://localhost:8321 --provider ollama --test-subdirs files`
2025-08-20 14:23:57 -04:00
..
_static feat: add batches API with OpenAI compatibility (with inference replay) (#3162) 2025-08-15 15:34:15 -07:00
notebooks chore: rename templates to distributions (#3035) 2025-08-04 11:34:17 -07:00
openapi_generator chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
resources Several documentation fixes and fix link to API reference 2025-02-04 14:00:43 -08:00
source feat: Add S3 Files Provider implementation 2025-08-20 14:23:57 -04:00
zero_to_hero_guide chore: rename templates to distributions (#3035) 2025-08-04 11:34:17 -07:00
conftest.py fix: sleep after notebook test 2025-03-23 14:03:35 -07:00
contbuild.sh Fix broken links with docs 2024-11-22 20:42:17 -08:00
dog.jpg Support for Llama3.2 models and Swift SDK (#98) 2024-09-25 10:29:58 -07:00
getting_started.ipynb chore: rename templates to distributions (#3035) 2025-08-04 11:34:17 -07:00
getting_started_llama4.ipynb chore: rename templates to distributions (#3035) 2025-08-04 11:34:17 -07:00
getting_started_llama_api.ipynb chore: rename templates to distributions (#3035) 2025-08-04 11:34:17 -07:00
license_header.txt Initial commit 2024-07-23 08:32:33 -07:00
make.bat feat(pre-commit): enhance pre-commit hooks with additional checks (#2014) 2025-04-30 11:35:49 -07:00
Makefile first version of readthedocs (#278) 2024-10-22 10:15:58 +05:30
original_rfc.md chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
quick_start.ipynb chore: rename templates to distributions (#3035) 2025-08-04 11:34:17 -07:00
README.md feat: add auto-generated CI documentation pre-commit hook (#2890) 2025-07-25 17:57:01 +02:00

Llama Stack Documentation

Here's a collection of comprehensive guides, examples, and resources for building AI applications with Llama Stack. For the complete documentation, visit our ReadTheDocs page.

Render locally

From the llama-stack root directory, run the following command to render the docs locally:

uv run --group docs sphinx-autobuild docs/source docs/build/html --write-all

You can open up the docs in your browser at http://localhost:8000

Content

Try out Llama Stack's capabilities through our detailed Jupyter notebooks: