diff --git a/llama_stack/scripts/generate_prompt_format.py b/llama_stack/scripts/generate_prompt_format.py index 338b23f3e..ec4c5e9be 100644 --- a/llama_stack/scripts/generate_prompt_format.py +++ b/llama_stack/scripts/generate_prompt_format.py @@ -18,7 +18,7 @@ import fire from llama_stack.models.llama.sku_list import resolve_model from llama_stack.providers.inline.inference.meta_reference.config import MetaReferenceInferenceConfig -from llama_stack.providers.inline.inference.meta_reference.generation import Llama +from llama_stack.providers.inline.inference.meta_reference.llama3.generation import Llama3 THIS_DIR = Path(__file__).parent.resolve() @@ -41,7 +41,7 @@ def run_main( llama_model = resolve_model(model_id) if not llama_model: raise ValueError(f"Model {model_id} not found") - generator = Llama.build( + generator = Llama3.build( config=config, model_id=model_id, llama_model=llama_model,