mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-02 08:44:44 +00:00
Merge branch 'main' into feat/litellm_sambanova_usage
This commit is contained in:
commit
d42a9ea53d
1 changed files with 0 additions and 3 deletions
|
@ -39,7 +39,6 @@ from llama_stack.apis.datasets import Datasets
|
||||||
from llama_stack.apis.post_training import (
|
from llama_stack.apis.post_training import (
|
||||||
Checkpoint,
|
Checkpoint,
|
||||||
DataConfig,
|
DataConfig,
|
||||||
EfficiencyConfig,
|
|
||||||
LoraFinetuningConfig,
|
LoraFinetuningConfig,
|
||||||
OptimizerConfig,
|
OptimizerConfig,
|
||||||
QATFinetuningConfig,
|
QATFinetuningConfig,
|
||||||
|
@ -90,8 +89,6 @@ class LoraFinetuningSingleDevice:
|
||||||
) -> None:
|
) -> None:
|
||||||
assert isinstance(training_config.data_config, DataConfig), "DataConfig must be initialized"
|
assert isinstance(training_config.data_config, DataConfig), "DataConfig must be initialized"
|
||||||
|
|
||||||
assert isinstance(training_config.efficiency_config, EfficiencyConfig), "EfficiencyConfig must be initialized"
|
|
||||||
|
|
||||||
self.job_uuid = job_uuid
|
self.job_uuid = job_uuid
|
||||||
self.training_config = training_config
|
self.training_config = training_config
|
||||||
if not isinstance(algorithm_config, LoraFinetuningConfig):
|
if not isinstance(algorithm_config, LoraFinetuningConfig):
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue