diff --git a/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md b/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md index 3fcb16e12..f532ce7ec 100644 --- a/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md +++ b/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md @@ -31,9 +31,9 @@ Note that you need access to nvidia GPUs to run this distribution. This distribu The following environment variables can be configured: - `LLAMASTACK_PORT`: Port for the Llama Stack distribution server (default: `5001`) -- `INFERENCE_MODEL`: Inference model loaded into the Meta Reference server (default: `Llama3.2-3B-Instruct`) +- `INFERENCE_MODEL`: Inference model loaded into the Meta Reference server (default: `meta-llama/Llama-3.2-3B-Instruct`) - `INFERENCE_CHECKPOINT_DIR`: Directory containing the Meta Reference model checkpoint (default: `null`) -- `SAFETY_MODEL`: Name of the safety (Llama-Guard) model to use (default: `Llama-Guard-3-1B`) +- `SAFETY_MODEL`: Name of the safety (Llama-Guard) model to use (default: `meta-llama/Llama-Guard-3-1B`) - `SAFETY_CHECKPOINT_DIR`: Directory containing the Llama-Guard model checkpoint (default: `null`) diff --git a/llama_stack/templates/meta-reference-gpu/meta_reference.py b/llama_stack/templates/meta-reference-gpu/meta_reference.py index 0c809016c..461d89a4a 100644 --- a/llama_stack/templates/meta-reference-gpu/meta_reference.py +++ b/llama_stack/templates/meta-reference-gpu/meta_reference.py @@ -112,7 +112,7 @@ def get_distribution_template() -> DistributionTemplate: "Port for the Llama Stack distribution server", ), "INFERENCE_MODEL": ( - "Llama3.2-3B-Instruct", + "meta-llama/Llama-3.2-3B-Instruct", "Inference model loaded into the Meta Reference server", ), "INFERENCE_CHECKPOINT_DIR": ( @@ -120,7 +120,7 @@ def get_distribution_template() -> DistributionTemplate: "Directory containing the Meta Reference model checkpoint", ), "SAFETY_MODEL": ( - "Llama-Guard-3-1B", + "meta-llama/Llama-Guard-3-1B", "Name of the safety (Llama-Guard) model to use", ), "SAFETY_CHECKPOINT_DIR": (