Revert provider / inference config back to mainline

This commit is contained in:
Connor Hack 2024-11-25 09:20:27 -08:00
parent 8d83759caf
commit bbea9bccf1
2 changed files with 3 additions and 3 deletions

View file

@ -36,8 +36,6 @@ class MetaReferenceInferenceImpl(Inference, ModelRegistryHelper, ModelsProtocolP
def __init__(self, config: MetaReferenceInferenceConfig) -> None: def __init__(self, config: MetaReferenceInferenceConfig) -> None:
self.config = config self.config = config
model = resolve_model(config.model) model = resolve_model(config.model)
if model is None:
raise RuntimeError(f"Unknown model: {config.model}, Run `llama model list`")
ModelRegistryHelper.__init__( ModelRegistryHelper.__init__(
self, self,
[ [
@ -47,6 +45,8 @@ class MetaReferenceInferenceImpl(Inference, ModelRegistryHelper, ModelsProtocolP
) )
], ],
) )
if model is None:
raise RuntimeError(f"Unknown model: {config.model}, Run `llama model list`")
self.model = model self.model = model
# verify that the checkpoint actually is for this model lol # verify that the checkpoint actually is for this model lol

View file

@ -81,7 +81,7 @@ def get_distribution_template() -> DistributionTemplate:
"Port for the Llama Stack distribution server", "Port for the Llama Stack distribution server",
), ),
"INFERENCE_MODEL": ( "INFERENCE_MODEL": (
"meta-llama/Llama3.2-3B-Instruct", "meta-llama/Llama-3.2-3B-Instruct",
"Inference model loaded into the Meta Reference server", "Inference model loaded into the Meta Reference server",
), ),
"INFERENCE_CHECKPOINT_DIR": ( "INFERENCE_CHECKPOINT_DIR": (