diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index d62186a47..dd429672f 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -74,8 +74,10 @@ pip install llama-stack-client Let's use the `llama-stack-client` CLI to check the connectivity to the server. ```bash -llama-stack-client configure --endpoint http://localhost:$LLAMA_STACK_PORT -llama-stack-client models list +$ llama-stack-client configure --endpoint http://localhost:$LLAMA_STACK_PORT +> Enter the API key (leave empty if no key is needed): +Done! You can now use the Llama Stack Client CLI with endpoint http://localhost:8321 +$ llama-stack-client models list ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━┓ ┃ identifier ┃ provider_id ┃ provider_resource_id ┃ metadata ┃ ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━┩