From e4c503d8f46ec8a2654f602642b0883c3a5f41ea Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Tue, 29 Oct 2024 14:19:49 -0700 Subject: [PATCH] content --- docs/source/getting_started/index.md | 12 +++++++++--- 1 file changed, 9 insertions(+), 3 deletions(-) diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index 882f8be52..564ed4b7c 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -8,7 +8,12 @@ To see more example apps built using Llama Stack, see [llama-stack-apps](https:/ ## Starting Up Llama Stack Server -### Decide your +```{toctree} +:maxdepth: 2 + +distributions/index +``` +### Decide Your Build Type There are two ways to start a Llama Stack: - **Docker**: we provide a number of pre-built Docker containers allowing you to get started instantly. If you are focused on application development, we recommend this option. @@ -43,6 +48,7 @@ If so, we suggest: - [Docker]() - [Conda]() + ## Testing with client Once the server is setup, we can test it with a client to see the example outputs by . This will run the chat completion client and query the distribution’s `/inference/chat_completion` API. Send a POST request to the server: @@ -73,8 +79,8 @@ You can find more example scripts with client SDKs to talk with the Llama Stack ```{toctree} -:maxdepth: 2 +:hidden: +:maxdepth: 1 developer_cookbook -distributions/index ```