temp commit

This commit is contained in:
Botao Chen 2024-11-27 14:35:01 -08:00
parent c31a78dfcb
commit 18ae577760
2 changed files with 21 additions and 20 deletions

View file

@ -183,16 +183,16 @@ class PostTraining(Protocol):
@webmethod(route="/post-training/supervised-fine-tune") @webmethod(route="/post-training/supervised-fine-tune")
def supervised_fine_tune( def supervised_fine_tune(
self, self,
job_uuid: str, job_uuid: Optional[str],
model: str, model: Optional[str],
dataset_id: str, dataset_id: Optional[str],
validation_dataset_id: str, validation_dataset_id: Optional[str],
algorithm: FinetuningAlgorithm, algorithm: Optional[FinetuningAlgorithm],
algorithm_config: LoraFinetuningConfig, algorithm_config: Optional[LoraFinetuningConfig],
optimizer_config: OptimizerConfig, optimizer_config: Optional[OptimizerConfig],
training_config: TrainingConfig, training_config: Optional[TrainingConfig],
hyperparam_search_config: Dict[str, Any], hyperparam_search_config: Optional[Dict[str, Any]],
logger_config: Dict[str, Any], logger_config: Optional[Dict[str, Any]],
) -> PostTrainingJob: ... ) -> PostTrainingJob: ...
@webmethod(route="/post-training/preference-optimize") @webmethod(route="/post-training/preference-optimize")

View file

@ -30,6 +30,7 @@ class MetaReferencePostTrainingImpl:
) )
OptimizerConfig( OptimizerConfig(
optimizer_type=OptimizerType.adamw,
lr=3e-4, lr=3e-4,
lr_min=3e-5, lr_min=3e-5,
weight_decay=0.1, weight_decay=0.1,
@ -50,16 +51,16 @@ class MetaReferencePostTrainingImpl:
def supervised_fine_tune( def supervised_fine_tune(
self, self,
job_uuid: str = "1234", job_uuid: Optional[str] = "1234",
model: str = " meta-llama/Llama-3.2-3B-Instruct", model: Optional[str] = " meta-llama/Llama-3.2-3B-Instruct",
dataset_id: str = "alpaca", dataset_id: Optional[str] = "alpaca",
validation_dataset_id: str = "alpaca", validation_dataset_id: Optional[str] = "alpaca",
algorithm: FinetuningAlgorithm = FinetuningAlgorithm.lora, algorithm: Optional[FinetuningAlgorithm] = FinetuningAlgorithm.lora,
algorithm_config: LoraFinetuningConfig = LoraFinetuningConfig, algorithm_config: Optional[LoraFinetuningConfig] = LoraFinetuningConfig,
optimizer_config: OptimizerConfig = OptimizerConfig, optimizer_config: Optional[OptimizerConfig] = OptimizerConfig,
training_config: TrainingConfig = TrainingConfig, training_config: Optional[TrainingConfig] = TrainingConfig,
hyperparam_search_config: Dict[str, Any] = {}, hyperparam_search_config: Optional[Dict[str, Any]] = {},
logger_config: Dict[str, Any] = {}, logger_config: Optional[Dict[str, Any]] = {},
) -> PostTrainingJob: ) -> PostTrainingJob:
# wrapper request to make it easier to pass around (internal only, not exposed to API) # wrapper request to make it easier to pass around (internal only, not exposed to API)
request = PostTrainingSFTRequest( request = PostTrainingSFTRequest(