From e178081be53e5047e3305cd6a694daff37c120bf Mon Sep 17 00:00:00 2001 From: Geronimo De Abreu Date: Fri, 8 Nov 2024 00:40:18 -0500 Subject: [PATCH] Use the list endpoint instead of ps to get ollama's models --- llama_stack/providers/remote/inference/ollama/ollama.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_stack/providers/remote/inference/ollama/ollama.py b/llama_stack/providers/remote/inference/ollama/ollama.py index 3530e1234..e57c6909f 100644 --- a/llama_stack/providers/remote/inference/ollama/ollama.py +++ b/llama_stack/providers/remote/inference/ollama/ollama.py @@ -72,7 +72,7 @@ class OllamaInferenceAdapter(Inference, ModelsProtocolPrivate): ollama_to_llama = {v: k for k, v in OLLAMA_SUPPORTED_MODELS.items()} ret = [] - res = await self.client.ps() + res = await self.client.list() for r in res["models"]: if r["model"] not in ollama_to_llama: print(f"Ollama is running a model unknown to Llama Stack: {r['model']}")