mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-30 23:51:00 +00:00
typo
This commit is contained in:
parent
7dc94b6b13
commit
499fe5ffe8
1 changed files with 7 additions and 7 deletions
|
@ -12,9 +12,9 @@ distributions/index
|
||||||
developer_cookbook
|
developer_cookbook
|
||||||
```
|
```
|
||||||
|
|
||||||
At the end of the guide, you will have learnt how to:
|
At the end of the guide, you will have learned how to:
|
||||||
- get a Llama Stack server up and running
|
- get a Llama Stack server up and running
|
||||||
- get a agent (with tool-calling, vector stores) which works with the above server
|
- set up an agent (with tool-calling and vector stores) that works with the above server
|
||||||
|
|
||||||
To see more example apps built using Llama Stack, see [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main).
|
To see more example apps built using Llama Stack, see [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main).
|
||||||
|
|
||||||
|
@ -48,7 +48,7 @@ If so, we suggest:
|
||||||
|
|
||||||
### Quick Start Commands
|
### Quick Start Commands
|
||||||
|
|
||||||
The following quick starts commands. Please visit each distribution page on detailed setup.
|
Once you have decided on the inference provider and distribution to use, use the following quick start commands to get started.
|
||||||
|
|
||||||
##### 1.0 Prerequisite
|
##### 1.0 Prerequisite
|
||||||
::::{tab-set}
|
::::{tab-set}
|
||||||
|
@ -237,7 +237,7 @@ inference:
|
||||||
- provider_id: ollama0
|
- provider_id: ollama0
|
||||||
provider_type: remote::ollama
|
provider_type: remote::ollama
|
||||||
config:
|
config:
|
||||||
url: http://127.0.0.1:14343
|
url: http://127.0.0.1:11434
|
||||||
```
|
```
|
||||||
|
|
||||||
```
|
```
|
||||||
|
@ -419,10 +419,10 @@ $ llama-stack-client models list
|
||||||
::::
|
::::
|
||||||
|
|
||||||
|
|
||||||
## Step 2. Build Your Llama Stack App
|
## Step 2. Run Llama Stack App
|
||||||
|
|
||||||
### chat_completion sanity test
|
### Chat Completion Test
|
||||||
Once the server is setup, we can test it with a client to see the example outputs by . This will run the chat completion client and query the distribution’s `/inference/chat_completion` API. Send a POST request to the server:
|
Once the server is set up, we can test it with a client to verify it's working correctly. The following command will send a chat completion request to the server's `/inference/chat_completion` API:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ curl http://localhost:5000/inference/chat_completion \
|
$ curl http://localhost:5000/inference/chat_completion \
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue