mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 12:07:34 +00:00
llama stack start -> llama stack run
This commit is contained in:
parent
fab6bd1728
commit
fb3c4566ce
5 changed files with 13 additions and 13 deletions
|
@ -242,7 +242,7 @@ An agentic app has several components including model inference, tool execution
|
|||
|
||||
The Llama Stack is a collection of REST APIs. An API is _implemented_ by Provider. An assembly of Providers together provides the implementation for the Stack -- this package is called a Distribution.
|
||||
|
||||
As an example, by running a simple command `llama stack start <YAML>`, you can bring up a server serving the following endpoints, among others:
|
||||
As an example, by running a simple command `llama stack run`, you can bring up a server serving the following endpoints, among others:
|
||||
```
|
||||
POST /inference/chat_completion
|
||||
POST /inference/completion
|
||||
|
@ -377,12 +377,12 @@ Now let’s start Llama Stack server.
|
|||
You need the YAML configuration file which was written out at the end by the `llama stack build` step.
|
||||
|
||||
```
|
||||
llama stack start local --name llama-8b --port 5000
|
||||
llama stack run local --name llama-8b --port 5000
|
||||
```
|
||||
You should see the Stack server start and print the APIs that it is supporting,
|
||||
|
||||
```
|
||||
$ llama stack start local --name llama-8b --port 5000
|
||||
$ llama stack run local --name llama-8b --port 5000
|
||||
|
||||
> initializing model parallel with size 1
|
||||
> initializing ddp with size 1
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue