mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-18 12:29:47 +00:00
fix misc
This commit is contained in:
parent
d55a8343ea
commit
d0a72cc288
3 changed files with 7 additions and 3 deletions
|
|
@ -39,8 +39,9 @@ class TorchtunePostTrainingImpl:
|
|||
checkpoint_dir: Optional[str],
|
||||
algorithm_config: Optional[AlgorithmConfig],
|
||||
) -> PostTrainingJob:
|
||||
if job_uuid in self.jobs_list:
|
||||
raise ValueError(f"Job {job_uuid} already exists")
|
||||
for job in self.jobs_list:
|
||||
if job_uuid == job.job_uuid:
|
||||
raise ValueError(f"Job {job_uuid} already exists")
|
||||
|
||||
post_training_job = PostTrainingJob(job_uuid=job_uuid)
|
||||
|
||||
|
|
|
|||
|
|
@ -19,6 +19,7 @@ class TestPostTraining:
|
|||
@pytest.mark.asyncio
|
||||
async def test_supervised_fine_tune(self, post_training_stack):
|
||||
algorithm_config = LoraFinetuningConfig(
|
||||
type="LoRA",
|
||||
lora_attn_modules=["q_proj", "v_proj", "output_proj"],
|
||||
apply_lora_to_mlp=True,
|
||||
apply_lora_to_output=False,
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue