mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 07:14:20 +00:00
docs: model name
This commit is contained in:
parent
dc822226cb
commit
a562d81825
1 changed files with 7 additions and 4 deletions
|
@ -13,7 +13,7 @@ Before running Llama Stack, you need to start the llama.cpp server:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# Example: Start llama.cpp server with a model
|
# Example: Start llama.cpp server with a model
|
||||||
./llama-server -m /path/to/your/model.gguf -c 4096 --host 0.0.0.0 --port 8080
|
./llama-server -m /path/to/your/YOUR_MODEL.gguf -c 4096 --host 0.0.0.0 --port 8080 --api-key YOUR_API_KEY --jinja -cb
|
||||||
```
|
```
|
||||||
|
|
||||||
Common llama.cpp server options:
|
Common llama.cpp server options:
|
||||||
|
@ -31,14 +31,17 @@ Set these environment variables before running Llama Stack:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export LLAMACPP_URL=http://localhost:8080 # URL of your llama.cpp server (without /v1 suffix)
|
export LLAMACPP_URL=http://localhost:8080 # URL of your llama.cpp server (without /v1 suffix)
|
||||||
export INFERENCE_MODEL=your-model-name # Name/identifier for your model
|
export INFERENCE_MODEL=your-model-name # Name/identifier without gguf extension
|
||||||
export LLAMACPP_API_KEY="" # API key (leave empty for local servers)
|
export LLAMACPP_API_KEY="YOUR_API_KEY" # API key (leave empty for local servers)
|
||||||
```
|
```
|
||||||
|
|
||||||
## Running Llama Stack
|
## Running Llama Stack
|
||||||
|
|
||||||
|
The model name will be you gguf file name without the extension.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
llama stack run llamacpp
|
llama stack build --template llamacpp --image-type conda
|
||||||
|
llama stack run llamacpp --image-type conda
|
||||||
```
|
```
|
||||||
|
|
||||||
## Configuration
|
## Configuration
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue