mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-06 02:32:40 +00:00
add comment about why ps vs list
This commit is contained in:
parent
4e81b1e650
commit
a49549063a
1 changed files with 3 additions and 0 deletions
|
@ -293,6 +293,9 @@ class OllamaInferenceAdapter(Inference, ModelsProtocolPrivate):
|
||||||
if model.model_type == ModelType.embedding:
|
if model.model_type == ModelType.embedding:
|
||||||
logger.info(f"Pulling embedding model `{model.provider_resource_id}` if necessary...")
|
logger.info(f"Pulling embedding model `{model.provider_resource_id}` if necessary...")
|
||||||
await self.client.pull(model.provider_resource_id)
|
await self.client.pull(model.provider_resource_id)
|
||||||
|
# we use list() here instead of ps() -
|
||||||
|
# - ps() only lists running models, not available models
|
||||||
|
# - models not currently running are run by the ollama server as needed
|
||||||
response = await self.client.list()
|
response = await self.client.list()
|
||||||
available_models = [m["model"] for m in response["models"]]
|
available_models = [m["model"] for m in response["models"]]
|
||||||
if model.provider_resource_id not in available_models:
|
if model.provider_resource_id not in available_models:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue