mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-05 10:13:05 +00:00
update docstring
This commit is contained in:
parent
d7ead08cb9
commit
d667a7109f
1 changed files with 28 additions and 0 deletions
|
@ -203,6 +203,34 @@ class NvidiaPostTrainingAdapter(ModelRegistryHelper):
|
|||
- NVIDIA_ACCESS_POLICIES: Access policies for the project
|
||||
- NVIDIA_DATASET_NAMESPACE: Namespace of the dataset
|
||||
- NVIDIA_OUTPUT_MODEL_DIR: Directory to save the output model
|
||||
|
||||
Supported models:
|
||||
- meta/llama-3.1-8b-instruct
|
||||
|
||||
Supported algorithm configs:
|
||||
- LoRA, SFT
|
||||
|
||||
Supported Parameters:
|
||||
- TrainingConfig:
|
||||
- n_epochs
|
||||
- data_config
|
||||
- optimizer_config
|
||||
- dtype
|
||||
- efficiency_config
|
||||
- max_steps_per_epoch
|
||||
- DataConfig:
|
||||
- dataset_id
|
||||
- batch_size
|
||||
- OptimizerConfig:
|
||||
- lr
|
||||
- LoRA config:
|
||||
- adapter_dim
|
||||
- adapter_dropout
|
||||
Note:
|
||||
- checkpoint_dir, hyperparam_search_config, logger_config are not supported atm, will be ignored
|
||||
- output_model_dir is set via environment variable NVIDIA_OUTPUT_MODEL_DIR
|
||||
|
||||
User is informed about unsupported parameters via warnings.
|
||||
"""
|
||||
# map model to nvidia model name
|
||||
nvidia_model = self.get_provider_model_id(model)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue