mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-27 06:28:50 +00:00
1.2 KiB
1.2 KiB
orphan |
---|
true |
inline::huggingface
Description
HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem.
Configuration
Field | Type | Required | Default | Description |
---|---|---|---|---|
device |
<class 'str'> |
No | cuda | |
distributed_backend |
Literal['fsdp', 'deepspeed' |
No | ||
checkpoint_format |
Literal['full_state', 'huggingface' |
No | huggingface | |
chat_template |
<class 'str'> |
No | ||
model_specific_config |
<class 'dict'> |
No | {'trust_remote_code': True, 'attn_implementation': 'sdpa'} | |
max_seq_length |
<class 'int'> |
No | 2048 | |
gradient_checkpointing |
<class 'bool'> |
No | False | |
save_total_limit |
<class 'int'> |
No | 3 | |
logging_steps |
<class 'int'> |
No | 10 | |
warmup_ratio |
<class 'float'> |
No | 0.1 | |
weight_decay |
<class 'float'> |
No | 0.01 | |
dataloader_num_workers |
<class 'int'> |
No | 4 | |
dataloader_pin_memory |
<class 'bool'> |
No | True |
Sample Configuration
checkpoint_format: huggingface
distributed_backend: null
device: cpu