diff --git a/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py b/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py index cc430577f..71b8bf759 100644 --- a/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py +++ b/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py @@ -43,7 +43,6 @@ from torchtune.modules.peft import ( get_adapter_state_dict, get_lora_module_names, get_merged_lora_ckpt, - load_dora_magnitudes, set_trainable_params, validate_missing_and_unexpected_for_lora, ) @@ -281,7 +280,6 @@ class LoraFinetuningSingleDevice: for m in model.modules(): if hasattr(m, "initialize_dora_magnitude"): m.initialize_dora_magnitude() - load_dora_magnitudes(model) if lora_weights_state_dict: lora_missing, lora_unexpected = model.load_state_dict( lora_weights_state_dict, strict=False