mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-03 09:21:45 +00:00
fix
This commit is contained in:
parent
06cb0c837e
commit
31b7e32ce6
1 changed files with 0 additions and 2 deletions
|
@ -43,7 +43,6 @@ from torchtune.modules.peft import (
|
||||||
get_adapter_state_dict,
|
get_adapter_state_dict,
|
||||||
get_lora_module_names,
|
get_lora_module_names,
|
||||||
get_merged_lora_ckpt,
|
get_merged_lora_ckpt,
|
||||||
load_dora_magnitudes,
|
|
||||||
set_trainable_params,
|
set_trainable_params,
|
||||||
validate_missing_and_unexpected_for_lora,
|
validate_missing_and_unexpected_for_lora,
|
||||||
)
|
)
|
||||||
|
@ -281,7 +280,6 @@ class LoraFinetuningSingleDevice:
|
||||||
for m in model.modules():
|
for m in model.modules():
|
||||||
if hasattr(m, "initialize_dora_magnitude"):
|
if hasattr(m, "initialize_dora_magnitude"):
|
||||||
m.initialize_dora_magnitude()
|
m.initialize_dora_magnitude()
|
||||||
load_dora_magnitudes(model)
|
|
||||||
if lora_weights_state_dict:
|
if lora_weights_state_dict:
|
||||||
lora_missing, lora_unexpected = model.load_state_dict(
|
lora_missing, lora_unexpected = model.load_state_dict(
|
||||||
lora_weights_state_dict, strict=False
|
lora_weights_state_dict, strict=False
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue