mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 22:39:47 +00:00
modify doc
This commit is contained in:
parent
486c0bc9c8
commit
85d0f5f528
7 changed files with 158 additions and 158 deletions
|
|
@ -53,7 +53,7 @@ docker run \
|
|||
-v ~/.llama:/root/.llama \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
|
@ -65,8 +65,8 @@ docker run \
|
|||
-v ~/.llama:/root/.llama \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=Llama-Guard-3-1B
|
||||
```
|
||||
|
||||
### Via Conda
|
||||
|
|
@ -77,7 +77,7 @@ Make sure you have done `pip install llama-stack` and have the Llama Stack CLI a
|
|||
llama stack build --template {{ name }} --image-type conda
|
||||
llama stack run distributions/{{ name }}/run.yaml \
|
||||
--port 5001 \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
|
@ -85,6 +85,6 @@ If you are using Llama Stack Safety / Shield APIs, use:
|
|||
```bash
|
||||
llama stack run distributions/{{ name }}/run-with-safety.yaml \
|
||||
--port 5001 \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=Llama-Guard-3-1B
|
||||
```
|
||||
|
|
|
|||
|
|
@ -112,7 +112,7 @@ def get_distribution_template() -> DistributionTemplate:
|
|||
"Port for the Llama Stack distribution server",
|
||||
),
|
||||
"INFERENCE_MODEL": (
|
||||
"meta-llama/Llama-3.2-3B-Instruct",
|
||||
"Llama3.2-3B-Instruct",
|
||||
"Inference model loaded into the Meta Reference server",
|
||||
),
|
||||
"INFERENCE_CHECKPOINT_DIR": (
|
||||
|
|
@ -120,7 +120,7 @@ def get_distribution_template() -> DistributionTemplate:
|
|||
"Directory containing the Meta Reference model checkpoint",
|
||||
),
|
||||
"SAFETY_MODEL": (
|
||||
"meta-llama/Llama-Guard-3-1B",
|
||||
"Llama-Guard-3-1B",
|
||||
"Name of the safety (Llama-Guard) model to use",
|
||||
),
|
||||
"SAFETY_CHECKPOINT_DIR": (
|
||||
|
|
|
|||
|
|
@ -55,7 +55,7 @@ docker run \
|
|||
-v ~/.llama:/root/.llama \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
|
@ -67,8 +67,8 @@ docker run \
|
|||
-v ~/.llama:/root/.llama \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=Llama-Guard-3-1B
|
||||
```
|
||||
|
||||
### Via Conda
|
||||
|
|
@ -79,7 +79,7 @@ Make sure you have done `pip install llama-stack` and have the Llama Stack CLI a
|
|||
llama stack build --template {{ name }} --image-type conda
|
||||
llama stack run distributions/{{ name }}/run.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
|
@ -87,6 +87,6 @@ If you are using Llama Stack Safety / Shield APIs, use:
|
|||
```bash
|
||||
llama stack run distributions/{{ name }}/run-with-safety.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=Llama-Guard-3-1B
|
||||
```
|
||||
|
|
|
|||
|
|
@ -84,7 +84,7 @@ def get_distribution_template() -> DistributionTemplate:
|
|||
"Port for the Llama Stack distribution server",
|
||||
),
|
||||
"INFERENCE_MODEL": (
|
||||
"meta-llama/Llama-3.2-3B-Instruct",
|
||||
"Llama3.2-3B-Instruct",
|
||||
"Inference model loaded into the Meta Reference server",
|
||||
),
|
||||
"INFERENCE_CHECKPOINT_DIR": (
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue