update docs

This commit is contained in:
Xi Yan 2024-11-08 12:46:43 -08:00
parent 5625aef48a
commit bd0622ef10

View file

@ -217,8 +217,7 @@ $ llama stack build --template meta-reference-gpu --image-type conda
3. Start running distribution 3. Start running distribution
``` ```
$ cd llama-stack/distributions/meta-reference-gpu $ llama stack run ~/.llama/distributions/llamastack-meta-reference-gpu/meta-reference-gpu-run.yaml
$ llama stack run ./run.yaml
``` ```
::: :::
@ -246,7 +245,7 @@ inference:
5. Start Llama Stack server 5. Start Llama Stack server
```bash ```bash
llama stack run ./gpu/run.yaml $ llama stack run ~/.llama/distributions/llamastack-tgi/tgi-run.yaml
``` ```
::: :::
@ -282,7 +281,7 @@ inference:
``` ```
llama stack build --template ollama --image-type conda llama stack build --template ollama --image-type conda
llama stack run ./gpu/run.yaml llama stack run ~/.llama/distributions/llamastack-ollama/ollama-run.yaml
``` ```
::: :::
@ -313,7 +312,7 @@ inference:
```bash ```bash
llama stack build --template together --image-type conda llama stack build --template together --image-type conda
# -- modify run.yaml to a valid Together server endpoint # -- modify run.yaml to a valid Together server endpoint
llama stack run ./run.yaml llama stack run ~/.llama/distributions/llamastack-together/together-run.yaml
``` ```
Make sure your `run.yaml` file has the inference provider pointing to the correct Together URL server endpoint. E.g. Make sure your `run.yaml` file has the inference provider pointing to the correct Together URL server endpoint. E.g.