From e93518fe065f230fdd0676a7600829b08c769252 Mon Sep 17 00:00:00 2001 From: sarthakdeshpande Date: Fri, 14 Mar 2025 17:18:20 +0530 Subject: [PATCH] fix: fixed import error --- llama_stack/scripts/generate_prompt_format.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llama_stack/scripts/generate_prompt_format.py b/llama_stack/scripts/generate_prompt_format.py index 338b23f3e..ec4c5e9be 100644 --- a/llama_stack/scripts/generate_prompt_format.py +++ b/llama_stack/scripts/generate_prompt_format.py @@ -18,7 +18,7 @@ import fire from llama_stack.models.llama.sku_list import resolve_model from llama_stack.providers.inline.inference.meta_reference.config import MetaReferenceInferenceConfig -from llama_stack.providers.inline.inference.meta_reference.generation import Llama +from llama_stack.providers.inline.inference.meta_reference.llama3.generation import Llama3 THIS_DIR = Path(__file__).parent.resolve() @@ -41,7 +41,7 @@ def run_main( llama_model = resolve_model(model_id) if not llama_model: raise ValueError(f"Model {model_id} not found") - generator = Llama.build( + generator = Llama3.build( config=config, model_id=model_id, llama_model=llama_model,