mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-02 16:54:42 +00:00
Revert provider / inference config back to mainline
This commit is contained in:
parent
8d83759caf
commit
bbea9bccf1
2 changed files with 3 additions and 3 deletions
|
@ -36,8 +36,6 @@ class MetaReferenceInferenceImpl(Inference, ModelRegistryHelper, ModelsProtocolP
|
||||||
def __init__(self, config: MetaReferenceInferenceConfig) -> None:
|
def __init__(self, config: MetaReferenceInferenceConfig) -> None:
|
||||||
self.config = config
|
self.config = config
|
||||||
model = resolve_model(config.model)
|
model = resolve_model(config.model)
|
||||||
if model is None:
|
|
||||||
raise RuntimeError(f"Unknown model: {config.model}, Run `llama model list`")
|
|
||||||
ModelRegistryHelper.__init__(
|
ModelRegistryHelper.__init__(
|
||||||
self,
|
self,
|
||||||
[
|
[
|
||||||
|
@ -47,6 +45,8 @@ class MetaReferenceInferenceImpl(Inference, ModelRegistryHelper, ModelsProtocolP
|
||||||
)
|
)
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
if model is None:
|
||||||
|
raise RuntimeError(f"Unknown model: {config.model}, Run `llama model list`")
|
||||||
self.model = model
|
self.model = model
|
||||||
# verify that the checkpoint actually is for this model lol
|
# verify that the checkpoint actually is for this model lol
|
||||||
|
|
||||||
|
|
|
@ -81,7 +81,7 @@ def get_distribution_template() -> DistributionTemplate:
|
||||||
"Port for the Llama Stack distribution server",
|
"Port for the Llama Stack distribution server",
|
||||||
),
|
),
|
||||||
"INFERENCE_MODEL": (
|
"INFERENCE_MODEL": (
|
||||||
"meta-llama/Llama3.2-3B-Instruct",
|
"meta-llama/Llama-3.2-3B-Instruct",
|
||||||
"Inference model loaded into the Meta Reference server",
|
"Inference model loaded into the Meta Reference server",
|
||||||
),
|
),
|
||||||
"INFERENCE_CHECKPOINT_DIR": (
|
"INFERENCE_CHECKPOINT_DIR": (
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue