diff --git a/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py b/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py index 7f1547657..cc430577f 100644 --- a/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py +++ b/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py @@ -110,6 +110,10 @@ class LoraFinetuningSingleDevice: self.checkpoint_dir = config.checkpoint_dir else: model = resolve_model(self.model_id) + if model is None: + raise ValueError( + f"{self.model_id} not found. Your model id should be in the llama models SKU list" + ) self.checkpoint_dir = model_checkpoint_dir(model) self._output_dir = str(DEFAULT_CHECKPOINT_DIR)