mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-30 07:39:38 +00:00
updates
This commit is contained in:
parent
f42f459f5d
commit
2c34f1eb52
1 changed files with 3 additions and 3 deletions
|
@ -39,7 +39,7 @@ There are two ways to start a Llama Stack:
|
||||||
|
|
||||||
Both of these provide options to run model inference using our reference implementations, Ollama, TGI, vLLM or even remote providers like Fireworks, Together, Bedrock, etc.
|
Both of these provide options to run model inference using our reference implementations, Ollama, TGI, vLLM or even remote providers like Fireworks, Together, Bedrock, etc.
|
||||||
|
|
||||||
1. **Starting up server via docker**:
|
### Docker
|
||||||
|
|
||||||
We provide pre-built Docker image of Llama Stack distribution, which can be found in the following links in the [distributions](../distributions/) folder.
|
We provide pre-built Docker image of Llama Stack distribution, which can be found in the following links in the [distributions](../distributions/) folder.
|
||||||
|
|
||||||
|
@ -68,9 +68,9 @@ docker run -it -p 5000:5000 -v ~/.llama:/root/.llama -v ./run.yaml:/root/my-run.
|
||||||
> Pro Tip: We may use `docker compose up` for starting up a distribution with remote providers (e.g. TGI) using [llamastack-local-cpu](https://hub.docker.com/repository/docker/llamastack/llamastack-local-cpu/general). You can checkout [these scripts](../distributions/) to help you get started.
|
> Pro Tip: We may use `docker compose up` for starting up a distribution with remote providers (e.g. TGI) using [llamastack-local-cpu](https://hub.docker.com/repository/docker/llamastack/llamastack-local-cpu/general). You can checkout [these scripts](../distributions/) to help you get started.
|
||||||
|
|
||||||
|
|
||||||
2. **Build->Configure->Run Llama Stack server via conda**:
|
### Conda
|
||||||
|
|
||||||
You may also build a LlamaStack distribution from scratch, configure it, and start running the distribution. This is useful for developing on LlamaStack.
|
You can use this method to build a Llama Stack distribution from scratch. This is useful when you intend to hack on the Llama Stack server codebase (or just want to understand.)
|
||||||
|
|
||||||
**`llama stack build`**
|
**`llama stack build`**
|
||||||
- You'll be prompted to enter build information interactively.
|
- You'll be prompted to enter build information interactively.
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue