fix: ensure ollama embedding model is registered properly in the template

This commit is contained in:
Ashwin Bharambe 2025-02-27 22:49:06 -08:00
parent 999195fe5b
commit 4c8a0fa8dc
3 changed files with 8 additions and 9 deletions

View file

@ -93,7 +93,7 @@ def get_distribution_template() -> DistributionTemplate:
"inference": [inference_provider],
"vector_io": [vector_io_provider_sqlite],
},
default_models=[inference_model],
default_models=[inference_model, embedding_model],
default_tool_groups=default_tool_groups,
),
"run-with-safety.yaml": RunConfigSettings(

View file

@ -90,6 +90,12 @@ models:
model_id: ${env.INFERENCE_MODEL}
provider_id: ollama
model_type: llm
- metadata:
embedding_dimension: 384
model_id: all-MiniLM-L6-v2
provider_id: ollama
provider_model_id: all-minilm:latest
model_type: embedding
shields: []
vector_dbs: []
datasets: []