From 7cbf3f8383ee3e9627935763e0f09479cdf856e3 Mon Sep 17 00:00:00 2001 From: Vladislav Date: Fri, 21 Feb 2025 13:56:16 +0100 Subject: [PATCH] fix after merge --- .../providers/remote/inference/groq/models.py | 17 +++++++---------- llama_stack/templates/groq/groq.py | 4 ++-- 2 files changed, 9 insertions(+), 12 deletions(-) diff --git a/llama_stack/providers/remote/inference/groq/models.py b/llama_stack/providers/remote/inference/groq/models.py index fd73e42e5..54ca2e839 100644 --- a/llama_stack/providers/remote/inference/groq/models.py +++ b/llama_stack/providers/remote/inference/groq/models.py @@ -5,25 +5,22 @@ # the root directory of this source tree. from llama_stack.models.llama.sku_list import CoreModelId -from llama_stack.providers.utils.inference.model_registry import ( - build_model_alias, - build_model_alias_with_just_provider_model_id, -) +from llama_stack.providers.utils.inference.model_registry import build_model_entry -_MODEL_ALIASES = [ - build_model_alias( +_MODEL_ENTRIES = [ + build_model_entry( "llama3-8b-8192", CoreModelId.llama3_1_8b_instruct.value, ), - build_model_alias_with_just_provider_model_id( + build_model_entry( "llama-3.1-8b-instant", CoreModelId.llama3_1_8b_instruct.value, ), - build_model_alias( + build_model_entry( "llama3-70b-8192", CoreModelId.llama3_70b_instruct.value, ), - build_model_alias( + build_model_entry( "llama-3.3-70b-versatile", CoreModelId.llama3_3_70b_instruct.value, ), @@ -31,7 +28,7 @@ _MODEL_ALIASES = [ # Preview models aren't recommended for production use, but we include this one # to pass the test fixture # TODO(aidand): Replace this with a stable model once Groq supports it - build_model_alias( + build_model_entry( "llama-3.2-3b-preview", CoreModelId.llama3_2_3b_instruct.value, ), diff --git a/llama_stack/templates/groq/groq.py b/llama_stack/templates/groq/groq.py index b81fc5e78..9e25f02cb 100644 --- a/llama_stack/templates/groq/groq.py +++ b/llama_stack/templates/groq/groq.py @@ -18,7 +18,7 @@ from llama_stack.providers.inline.inference.sentence_transformers import ( ) from llama_stack.providers.inline.vector_io.faiss.config import FaissVectorIOConfig from llama_stack.providers.remote.inference.groq import GroqConfig -from llama_stack.providers.remote.inference.groq.models import _MODEL_ALIASES +from llama_stack.providers.remote.inference.groq.models import _MODEL_ENTRIES from llama_stack.templates.template import DistributionTemplate, RunConfigSettings @@ -73,7 +73,7 @@ def get_distribution_template() -> DistributionTemplate: provider_model_id=m.provider_model_id, provider_id=name, ) - for m in _MODEL_ALIASES + for m in _MODEL_ENTRIES ] default_tool_groups = [