mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-02 00:34:44 +00:00
temp commit
This commit is contained in:
parent
c31a78dfcb
commit
18ae577760
2 changed files with 21 additions and 20 deletions
|
@ -183,16 +183,16 @@ class PostTraining(Protocol):
|
||||||
@webmethod(route="/post-training/supervised-fine-tune")
|
@webmethod(route="/post-training/supervised-fine-tune")
|
||||||
def supervised_fine_tune(
|
def supervised_fine_tune(
|
||||||
self,
|
self,
|
||||||
job_uuid: str,
|
job_uuid: Optional[str],
|
||||||
model: str,
|
model: Optional[str],
|
||||||
dataset_id: str,
|
dataset_id: Optional[str],
|
||||||
validation_dataset_id: str,
|
validation_dataset_id: Optional[str],
|
||||||
algorithm: FinetuningAlgorithm,
|
algorithm: Optional[FinetuningAlgorithm],
|
||||||
algorithm_config: LoraFinetuningConfig,
|
algorithm_config: Optional[LoraFinetuningConfig],
|
||||||
optimizer_config: OptimizerConfig,
|
optimizer_config: Optional[OptimizerConfig],
|
||||||
training_config: TrainingConfig,
|
training_config: Optional[TrainingConfig],
|
||||||
hyperparam_search_config: Dict[str, Any],
|
hyperparam_search_config: Optional[Dict[str, Any]],
|
||||||
logger_config: Dict[str, Any],
|
logger_config: Optional[Dict[str, Any]],
|
||||||
) -> PostTrainingJob: ...
|
) -> PostTrainingJob: ...
|
||||||
|
|
||||||
@webmethod(route="/post-training/preference-optimize")
|
@webmethod(route="/post-training/preference-optimize")
|
||||||
|
|
|
@ -30,6 +30,7 @@ class MetaReferencePostTrainingImpl:
|
||||||
)
|
)
|
||||||
|
|
||||||
OptimizerConfig(
|
OptimizerConfig(
|
||||||
|
optimizer_type=OptimizerType.adamw,
|
||||||
lr=3e-4,
|
lr=3e-4,
|
||||||
lr_min=3e-5,
|
lr_min=3e-5,
|
||||||
weight_decay=0.1,
|
weight_decay=0.1,
|
||||||
|
@ -50,16 +51,16 @@ class MetaReferencePostTrainingImpl:
|
||||||
|
|
||||||
def supervised_fine_tune(
|
def supervised_fine_tune(
|
||||||
self,
|
self,
|
||||||
job_uuid: str = "1234",
|
job_uuid: Optional[str] = "1234",
|
||||||
model: str = " meta-llama/Llama-3.2-3B-Instruct",
|
model: Optional[str] = " meta-llama/Llama-3.2-3B-Instruct",
|
||||||
dataset_id: str = "alpaca",
|
dataset_id: Optional[str] = "alpaca",
|
||||||
validation_dataset_id: str = "alpaca",
|
validation_dataset_id: Optional[str] = "alpaca",
|
||||||
algorithm: FinetuningAlgorithm = FinetuningAlgorithm.lora,
|
algorithm: Optional[FinetuningAlgorithm] = FinetuningAlgorithm.lora,
|
||||||
algorithm_config: LoraFinetuningConfig = LoraFinetuningConfig,
|
algorithm_config: Optional[LoraFinetuningConfig] = LoraFinetuningConfig,
|
||||||
optimizer_config: OptimizerConfig = OptimizerConfig,
|
optimizer_config: Optional[OptimizerConfig] = OptimizerConfig,
|
||||||
training_config: TrainingConfig = TrainingConfig,
|
training_config: Optional[TrainingConfig] = TrainingConfig,
|
||||||
hyperparam_search_config: Dict[str, Any] = {},
|
hyperparam_search_config: Optional[Dict[str, Any]] = {},
|
||||||
logger_config: Dict[str, Any] = {},
|
logger_config: Optional[Dict[str, Any]] = {},
|
||||||
) -> PostTrainingJob:
|
) -> PostTrainingJob:
|
||||||
# wrapper request to make it easier to pass around (internal only, not exposed to API)
|
# wrapper request to make it easier to pass around (internal only, not exposed to API)
|
||||||
request = PostTrainingSFTRequest(
|
request = PostTrainingSFTRequest(
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue