From e8d3eee0954737b0e247842fc650dd3d4677cc2a Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Mon, 18 Nov 2024 23:51:25 -0800 Subject: [PATCH] Fix docs yet again --- .../distributions/self_hosted_distro/fireworks.md | 2 +- .../distributions/self_hosted_distro/remote-vllm.md | 4 ++-- .../getting_started/distributions/self_hosted_distro/tgi.md | 4 ++-- .../distributions/self_hosted_distro/together.md | 2 +- llama_stack/templates/fireworks/doc_template.md | 2 +- llama_stack/templates/remote-vllm/doc_template.md | 4 ++-- llama_stack/templates/tgi/doc_template.md | 4 ++-- llama_stack/templates/together/doc_template.md | 2 +- 8 files changed, 12 insertions(+), 12 deletions(-) diff --git a/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md b/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md index 30d822946..f940e6de2 100644 --- a/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md +++ b/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md @@ -53,7 +53,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-fireworks \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env FIREWORKS_API_KEY=$FIREWORKS_API_KEY ``` diff --git a/docs/source/getting_started/distributions/self_hosted_distro/remote-vllm.md b/docs/source/getting_started/distributions/self_hosted_distro/remote-vllm.md index 884e9a13c..748b98732 100644 --- a/docs/source/getting_started/distributions/self_hosted_distro/remote-vllm.md +++ b/docs/source/getting_started/distributions/self_hosted_distro/remote-vllm.md @@ -85,7 +85,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-remote-vllm \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env VLLM_URL=http://host.docker.internal:$INFERENCE_PORT/v1 @@ -102,7 +102,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run-with-safety.yaml:/root/my-run.yaml \ llamastack/distribution-remote-vllm \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env VLLM_URL=http://host.docker.internal:$INFERENCE_PORT/v1 \ diff --git a/docs/source/getting_started/distributions/self_hosted_distro/tgi.md b/docs/source/getting_started/distributions/self_hosted_distro/tgi.md index 7f84833f3..63631f937 100644 --- a/docs/source/getting_started/distributions/self_hosted_distro/tgi.md +++ b/docs/source/getting_started/distributions/self_hosted_distro/tgi.md @@ -80,7 +80,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-tgi \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env TGI_URL=http://host.docker.internal:$INFERENCE_PORT @@ -94,7 +94,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run-with-safety.yaml:/root/my-run.yaml \ llamastack/distribution-tgi \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env TGI_URL=http://host.docker.internal:$INFERENCE_PORT \ diff --git a/docs/source/getting_started/distributions/self_hosted_distro/together.md b/docs/source/getting_started/distributions/self_hosted_distro/together.md index fe4dc5fed..5d79fcf0c 100644 --- a/docs/source/getting_started/distributions/self_hosted_distro/together.md +++ b/docs/source/getting_started/distributions/self_hosted_distro/together.md @@ -52,7 +52,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-together \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env TOGETHER_API_KEY=$TOGETHER_API_KEY ``` diff --git a/llama_stack/templates/fireworks/doc_template.md b/llama_stack/templates/fireworks/doc_template.md index 6f6da3a91..2a91ece07 100644 --- a/llama_stack/templates/fireworks/doc_template.md +++ b/llama_stack/templates/fireworks/doc_template.md @@ -45,7 +45,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-{{ name }} \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env FIREWORKS_API_KEY=$FIREWORKS_API_KEY ``` diff --git a/llama_stack/templates/remote-vllm/doc_template.md b/llama_stack/templates/remote-vllm/doc_template.md index aca4fc643..63432fb70 100644 --- a/llama_stack/templates/remote-vllm/doc_template.md +++ b/llama_stack/templates/remote-vllm/doc_template.md @@ -77,7 +77,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-{{ name }} \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env VLLM_URL=http://host.docker.internal:$INFERENCE_PORT/v1 @@ -94,7 +94,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run-with-safety.yaml:/root/my-run.yaml \ llamastack/distribution-{{ name }} \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env VLLM_URL=http://host.docker.internal:$INFERENCE_PORT/v1 \ diff --git a/llama_stack/templates/tgi/doc_template.md b/llama_stack/templates/tgi/doc_template.md index d4dee7fb7..0f6001e1a 100644 --- a/llama_stack/templates/tgi/doc_template.md +++ b/llama_stack/templates/tgi/doc_template.md @@ -73,7 +73,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-{{ name }} \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env TGI_URL=http://host.docker.internal:$INFERENCE_PORT @@ -87,7 +87,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run-with-safety.yaml:/root/my-run.yaml \ llamastack/distribution-{{ name }} \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env INFERENCE_MODEL=$INFERENCE_MODEL \ --env TGI_URL=http://host.docker.internal:$INFERENCE_PORT \ diff --git a/llama_stack/templates/together/doc_template.md b/llama_stack/templates/together/doc_template.md index 667a68713..5c1580dac 100644 --- a/llama_stack/templates/together/doc_template.md +++ b/llama_stack/templates/together/doc_template.md @@ -45,7 +45,7 @@ docker run \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -v ./run.yaml:/root/my-run.yaml \ llamastack/distribution-{{ name }} \ - /root/my-run.yaml \ + --yaml-config /root/my-run.yaml \ --port $LLAMA_STACK_PORT \ --env TOGETHER_API_KEY=$TOGETHER_API_KEY ```