mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-01 00:05:18 +00:00
fix: Update SFTConfig parameter to fix CI and Post Training Workflow (#2948)
# What does this PR do? - Change max_seq_length to max_length in SFTConfig constructor - TRL deprecated max_seq_length in Feb 2024 and removed it in v0.20.0 - Reference: https://github.com/huggingface/trl/pull/2895 This resolves the SFT training failure in CI tests
This commit is contained in:
parent
c7dc0f21b4
commit
58ffd82853
1 changed files with 1 additions and 1 deletions
|
@ -469,7 +469,7 @@ class HFFinetuningSingleDevice:
|
||||||
use_cpu=True if device.type == "cpu" and not torch.backends.mps.is_available() else False,
|
use_cpu=True if device.type == "cpu" and not torch.backends.mps.is_available() else False,
|
||||||
save_strategy=save_strategy,
|
save_strategy=save_strategy,
|
||||||
report_to="none",
|
report_to="none",
|
||||||
max_seq_length=provider_config.max_seq_length,
|
max_length=provider_config.max_seq_length,
|
||||||
gradient_accumulation_steps=config.gradient_accumulation_steps,
|
gradient_accumulation_steps=config.gradient_accumulation_steps,
|
||||||
gradient_checkpointing=provider_config.gradient_checkpointing,
|
gradient_checkpointing=provider_config.gradient_checkpointing,
|
||||||
learning_rate=lr,
|
learning_rate=lr,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue