From 28ce51198681c2f5b1c1d0a5a0f61f96e7b5d260 Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Thu, 26 Dec 2024 14:32:07 -0800 Subject: [PATCH] fix --endpoint docs --- docs/source/getting_started/index.md | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index c6227db99..80590bfad 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -51,7 +51,8 @@ pip install llama-stack-client Let's use the `llama-stack-client` CLI to check the connectivity to the server. ```bash -llama-stack-client --endpoint http://localhost:$LLAMA_STACK_PORT models list +llama-stack-client configure --endpoint http://localhost:$LLAMA_STACK_PORT +llama-stack-client models list ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━┓ ┃ identifier ┃ provider_id ┃ provider_resource_id ┃ metadata ┃ ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━┩ @@ -61,7 +62,7 @@ llama-stack-client --endpoint http://localhost:$LLAMA_STACK_PORT models list You can test basic Llama inference completion using the CLI too. ```bash -llama-stack-client --endpoint http://localhost:$LLAMA_STACK_PORT \ +llama-stack-client inference chat-completion \ --message "hello, what model are you?" ```