mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-16 06:27:58 +00:00
**Description** This PR removes some of the warnings when uv builds the docs - Errors appear when generating docs about .md files not appearing in toctree. ~~Adding content to the `providers-gen.py ` file that adds `--- orphan: true ---` to to each file.~~. Added a toctree generator to the `providers-gen.py` file, this gets rid of the errors in the builds. - Deletes the `_openai_compat` files, extension of PR #2849 - Adds the `files` APIs section to the `providers` toctree on the index page - Manually adds the `--- orphan: true ---` to the advanced apis. Ill try to find a way to modify the providers code gen so it automatically adds it, but this fixes the errors. - Adds the `testing.md` to the `contributing` toctree - Adds `starting_llama_stack_server.md` to `distributions` toctree There are some other warnings im still looking at but this PR gets rid of most of the toctree errors Theres also an issue with the actual distribution-codegen that I can investigate in another PR. Opened a bug for it here #2873
1.2 KiB
1.2 KiB
orphan |
---|
true |
inline::huggingface
Description
HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem.
Configuration
Field | Type | Required | Default | Description |
---|---|---|---|---|
device |
<class 'str'> |
No | cuda | |
distributed_backend |
Literal['fsdp', 'deepspeed' |
No | ||
checkpoint_format |
Literal['full_state', 'huggingface' |
No | huggingface | |
chat_template |
<class 'str'> |
No | ||
model_specific_config |
<class 'dict'> |
No | {'trust_remote_code': True, 'attn_implementation': 'sdpa'} | |
max_seq_length |
<class 'int'> |
No | 2048 | |
gradient_checkpointing |
<class 'bool'> |
No | False | |
save_total_limit |
<class 'int'> |
No | 3 | |
logging_steps |
<class 'int'> |
No | 10 | |
warmup_ratio |
<class 'float'> |
No | 0.1 | |
weight_decay |
<class 'float'> |
No | 0.01 | |
dataloader_num_workers |
<class 'int'> |
No | 4 | |
dataloader_pin_memory |
<class 'bool'> |
No | True |
Sample Configuration
checkpoint_format: huggingface
distributed_backend: null
device: cpu