mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-18 01:27:15 +00:00
add embedding model by default
This commit is contained in:
parent
2a9b13dd52
commit
2f88006bd0
43 changed files with 446 additions and 85 deletions
|
|
@ -337,7 +337,7 @@ class OllamaInferenceAdapter(Inference, ModelsProtocolPrivate):
|
|||
|
||||
async def register_model(self, model: Model) -> Model:
|
||||
# ollama does not have embedding models running. Check if the model is in list of available models.
|
||||
if model.model_type == ModelType.embedding_model:
|
||||
if model.model_type == ModelType.embedding:
|
||||
response = await self.client.list()
|
||||
available_models = [m["model"] for m in response["models"]]
|
||||
if model.provider_resource_id not in available_models:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue