mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-31 16:01:46 +00:00
Use the list endpoint instead of ps to get ollama's models
This commit is contained in:
parent
6192bf43a4
commit
e178081be5
1 changed files with 1 additions and 1 deletions
|
@ -72,7 +72,7 @@ class OllamaInferenceAdapter(Inference, ModelsProtocolPrivate):
|
||||||
ollama_to_llama = {v: k for k, v in OLLAMA_SUPPORTED_MODELS.items()}
|
ollama_to_llama = {v: k for k, v in OLLAMA_SUPPORTED_MODELS.items()}
|
||||||
|
|
||||||
ret = []
|
ret = []
|
||||||
res = await self.client.ps()
|
res = await self.client.list()
|
||||||
for r in res["models"]:
|
for r in res["models"]:
|
||||||
if r["model"] not in ollama_to_llama:
|
if r["model"] not in ollama_to_llama:
|
||||||
print(f"Ollama is running a model unknown to Llama Stack: {r['model']}")
|
print(f"Ollama is running a model unknown to Llama Stack: {r['model']}")
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue