From d04c9d9e4e583051a77542e4740d53571d09d254 Mon Sep 17 00:00:00 2001 From: Dinesh Yeduguru Date: Mon, 18 Nov 2024 23:43:47 -0800 Subject: [PATCH] change function name --- .../providers/remote/inference/ollama/ollama.py | 16 ++++++++-------- .../providers/utils/inference/model_registry.py | 2 +- 2 files changed, 9 insertions(+), 9 deletions(-) diff --git a/llama_stack/providers/remote/inference/ollama/ollama.py b/llama_stack/providers/remote/inference/ollama/ollama.py index f06d9fad7..33670d6de 100644 --- a/llama_stack/providers/remote/inference/ollama/ollama.py +++ b/llama_stack/providers/remote/inference/ollama/ollama.py @@ -16,7 +16,7 @@ from ollama import AsyncClient from llama_stack.providers.utils.inference.model_registry import ( build_model_alias, - build_model_alias_with_just_llama_model, + build_model_alias_with_just_provider_model_id, ModelRegistryHelper, ) @@ -45,7 +45,7 @@ model_aliases = [ "llama3.1:8b-instruct-fp16", CoreModelId.llama3_1_8b_instruct.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama3.1:8b", CoreModelId.llama3_1_8b_instruct.value, ), @@ -53,7 +53,7 @@ model_aliases = [ "llama3.1:70b-instruct-fp16", CoreModelId.llama3_1_70b_instruct.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama3.1:70b", CoreModelId.llama3_1_70b_instruct.value, ), @@ -65,19 +65,19 @@ model_aliases = [ "llama3.2:3b-instruct-fp16", CoreModelId.llama3_2_3b_instruct.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama3.2:1b", CoreModelId.llama3_2_1b_instruct.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama3.2:3b", CoreModelId.llama3_2_3b_instruct.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama-guard3:8b", CoreModelId.llama_guard_3_8b.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama-guard3:1b", CoreModelId.llama_guard_3_1b.value, ), @@ -85,7 +85,7 @@ model_aliases = [ "x/llama3.2-vision:11b-instruct-fp16", CoreModelId.llama3_2_11b_vision_instruct.value, ), - build_model_alias_with_just_llama_model( + build_model_alias_with_just_provider_model_id( "llama3.2-vision", CoreModelId.llama3_2_11b_vision_instruct.value, ), diff --git a/llama_stack/providers/utils/inference/model_registry.py b/llama_stack/providers/utils/inference/model_registry.py index 5ac738f47..07225fac0 100644 --- a/llama_stack/providers/utils/inference/model_registry.py +++ b/llama_stack/providers/utils/inference/model_registry.py @@ -36,7 +36,7 @@ def build_model_alias(provider_model_id: str, model_descriptor: str) -> ModelAli ) -def build_model_alias_with_just_llama_model( +def build_model_alias_with_just_provider_model_id( provider_model_id: str, model_descriptor: str ) -> ModelAlias: return ModelAlias(