From bd0622ef104c2edd78cbf91c992ed545b89bb636 Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Fri, 8 Nov 2024 12:46:43 -0800 Subject: [PATCH] update docs --- docs/source/getting_started/index.md | 9 ++++----- 1 file changed, 4 insertions(+), 5 deletions(-) diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index 31a6fc026..92643d87e 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -217,8 +217,7 @@ $ llama stack build --template meta-reference-gpu --image-type conda 3. Start running distribution ``` -$ cd llama-stack/distributions/meta-reference-gpu -$ llama stack run ./run.yaml +$ llama stack run ~/.llama/distributions/llamastack-meta-reference-gpu/meta-reference-gpu-run.yaml ``` ::: @@ -246,7 +245,7 @@ inference: 5. Start Llama Stack server ```bash -llama stack run ./gpu/run.yaml +$ llama stack run ~/.llama/distributions/llamastack-tgi/tgi-run.yaml ``` ::: @@ -282,7 +281,7 @@ inference: ``` llama stack build --template ollama --image-type conda -llama stack run ./gpu/run.yaml +llama stack run ~/.llama/distributions/llamastack-ollama/ollama-run.yaml ``` ::: @@ -313,7 +312,7 @@ inference: ```bash llama stack build --template together --image-type conda # -- modify run.yaml to a valid Together server endpoint -llama stack run ./run.yaml +llama stack run ~/.llama/distributions/llamastack-together/together-run.yaml ``` Make sure your `run.yaml` file has the inference provider pointing to the correct Together URL server endpoint. E.g.