From c590674ee2f44eef83b796eea9fe3183a920b733 Mon Sep 17 00:00:00 2001 From: Matthew Farrellee Date: Sun, 27 Apr 2025 10:44:45 -0400 Subject: [PATCH] live listing overrides static listing for ollama & vllm model registration --- llama_stack/providers/remote/inference/ollama/ollama.py | 5 ++++- llama_stack/providers/remote/inference/vllm/vllm.py | 5 ++++- 2 files changed, 8 insertions(+), 2 deletions(-) diff --git a/llama_stack/providers/remote/inference/ollama/ollama.py b/llama_stack/providers/remote/inference/ollama/ollama.py index 481c56fe5..2282e2726 100644 --- a/llama_stack/providers/remote/inference/ollama/ollama.py +++ b/llama_stack/providers/remote/inference/ollama/ollama.py @@ -333,7 +333,10 @@ class OllamaInferenceAdapter( return EmbeddingsResponse(embeddings=embeddings) async def register_model(self, model: Model) -> Model: - model = await self.register_helper.register_model(model) + try: + model = await self.register_helper.register_model(model) + except ValueError: + pass # Ignore statically unknown model, will check live listing if model.model_type == ModelType.embedding: logger.info(f"Pulling embedding model `{model.provider_resource_id}` if necessary...") await self.client.pull(model.provider_resource_id) diff --git a/llama_stack/providers/remote/inference/vllm/vllm.py b/llama_stack/providers/remote/inference/vllm/vllm.py index d141afa86..4d3aafd6a 100644 --- a/llama_stack/providers/remote/inference/vllm/vllm.py +++ b/llama_stack/providers/remote/inference/vllm/vllm.py @@ -358,7 +358,10 @@ class VLLMInferenceAdapter(Inference, ModelsProtocolPrivate): async def register_model(self, model: Model) -> Model: assert self.client is not None - model = await self.register_helper.register_model(model) + try: + model = await self.register_helper.register_model(model) + except ValueError: + pass # Ignore statically unknown model, will check live listing res = await self.client.models.list() available_models = [m.id async for m in res] if model.provider_resource_id not in available_models: