llama-stack/llama_stack/providers/inline/post_training/torchtune
Charlie Doern 0751a960a5
feat: make training config fields optional (#1861)
# What does this PR do?

Today, supervised_fine_tune itself and the `TrainingConfig` class have a
bunch of required fields that a provider implementation might not need.

for example, if a provider wants to handle hyperparameters in its
configuration as well as any type of dataset retrieval, optimizer or
LoRA config, a user will still need to pass in a virtually empty
`DataConfig`, `OptimizerConfig` and `AlgorithmConfig` in some cases.

Many of these fields are intended to work specifically with llama models
and knobs intended for customizing inline.

Adding remote post_training providers will require loosening these
arguments, or forcing users to pass in empty objects to satisfy the
pydantic models.

Signed-off-by: Charlie Doern <cdoern@redhat.com>
2025-04-12 01:13:45 -07:00
..
common refactor: move all llama code to models/llama out of meta reference (#1887) 2025-04-07 15:03:58 -07:00
datasets chore: fix mypy violations in post_training modules (#1548) 2025-03-18 14:58:16 -07:00
recipes feat: make training config fields optional (#1861) 2025-04-12 01:13:45 -07:00
__init__.py chore: fix typing hints for get_provider_impl deps arguments (#1544) 2025-03-11 10:07:28 -07:00
config.py test: add unit test to ensure all config types are instantiable (#1601) 2025-03-12 22:29:58 -07:00
post_training.py refactor: move all datetime.now() calls to UTC (#1589) 2025-03-13 15:34:53 -07:00