mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-23 22:03:55 +00:00
docs: clarify run.yaml files are starting points for customization
- Add new documentation section on customizing run.yaml files - Clarify that generated run.yaml files are templates, not production configs - Add guidance on customization best practices and common scenarios - Update existing documentation to reference customization guide - Improve clarity around run.yaml file usage for better user experience This enhancement makes it clearer for users that run.yaml files are meant to be customized for production use, improving the overall developer experience.
This commit is contained in:
parent
618ccea090
commit
9613c63ca6
5 changed files with 50 additions and 1 deletions
|
|
@ -54,7 +54,7 @@ Llama Stack is a server that exposes multiple APIs, you connect with it using th
|
|||
You can use Python to build and run the Llama Stack server, which is useful for testing and development.
|
||||
|
||||
Llama Stack uses a [YAML configuration file](../distributions/configuration.md) to specify the stack setup,
|
||||
which defines the providers and their settings.
|
||||
which defines the providers and their settings. The generated configuration serves as a starting point that you can [customize for your specific needs](../distributions/customizing_run_yaml.md).
|
||||
Now let's build and run the Llama Stack config for Ollama.
|
||||
We use `starter` as template. By default all providers are disabled, this requires enable ollama by passing environment variables.
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue