mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-27 06:28:50 +00:00
1.4 KiB
1.4 KiB
inline::huggingface
Description
HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem.
Configuration
Field | Type | Required | Default | Description |
---|---|---|---|---|
device |
<class 'str'> |
No | cuda | |
distributed_backend |
Literal['fsdp', 'deepspeed' |
No | ||
checkpoint_format |
Literal['full_state', 'huggingface' |
No | huggingface | |
chat_template |
<class 'str'> |
No | < | user |
{input} | ||||
< | assistant | > | ||
{output} | ||||
model_specific_config |
<class 'dict'> |
No | {'trust_remote_code': True, 'attn_implementation': 'sdpa'} | |
max_seq_length |
<class 'int'> |
No | 2048 | |
gradient_checkpointing |
<class 'bool'> |
No | False | |
save_total_limit |
<class 'int'> |
No | 3 | |
logging_steps |
<class 'int'> |
No | 10 | |
warmup_ratio |
<class 'float'> |
No | 0.1 | |
weight_decay |
<class 'float'> |
No | 0.01 | |
dataloader_num_workers |
<class 'int'> |
No | 4 | |
dataloader_pin_memory |
<class 'bool'> |
No | True | |
dpo_beta |
<class 'float'> |
No | 0.1 | |
use_reference_model |
<class 'bool'> |
No | True | |
dpo_loss_type |
Literal['sigmoid', 'hinge', 'ipo', 'kto_pair' |
No | sigmoid |
Sample Configuration
checkpoint_format: huggingface
distributed_backend: null
device: cpu