llama-stack/llama_stack/providers/inline/post_training
Charlie Doern 0751a960a5
feat: make training config fields optional (#1861)
# What does this PR do?

Today, supervised_fine_tune itself and the `TrainingConfig` class have a
bunch of required fields that a provider implementation might not need.

for example, if a provider wants to handle hyperparameters in its
configuration as well as any type of dataset retrieval, optimizer or
LoRA config, a user will still need to pass in a virtually empty
`DataConfig`, `OptimizerConfig` and `AlgorithmConfig` in some cases.

Many of these fields are intended to work specifically with llama models
and knobs intended for customizing inline.

Adding remote post_training providers will require loosening these
arguments, or forcing users to pass in empty objects to satisfy the
pydantic models.

Signed-off-by: Charlie Doern <cdoern@redhat.com>
2025-04-12 01:13:45 -07:00
..
common chore: fix mypy violations in post_training modules (#1548) 2025-03-18 14:58:16 -07:00
torchtune feat: make training config fields optional (#1861) 2025-04-12 01:13:45 -07:00
__init__.py Add init files to post training folders (#711) 2025-01-13 20:19:18 -08:00