refactor: set proper name for embedding all-minilm:l6-v2 model

- we are using all-minilm:l6-v2 but the model we download from ollama is all-minilm:latest
- even currently they are exactly the same model but if l12-v2 is updated, "latest" might not still valid.
- the only change in this PR is pin the model from ollama

Signed-off-by: Wen Zhou <wenzhou@redhat.com>
This commit is contained in:
Wen Zhou 2025-07-04 15:53:53 +02:00
parent ea966565f6
commit 214b1fe1ae
4 changed files with 5 additions and 5 deletions

View file

@ -180,9 +180,9 @@ Available Models
┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━┓ ┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━┓
┃ model_type ┃ identifier ┃ provider_resource_id ┃ metadata ┃ provider_id ┃ ┃ model_type ┃ identifier ┃ provider_resource_id ┃ metadata ┃ provider_id ┃
┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━┩ ┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━┩
│ embedding │ all-MiniLM-L6-v2 │ all-minilm:latest │ {'embedding_dimension': 384.0} │ ollama │ │ embedding │ all-MiniLM-L6-v2 │ all-minilm:l6-v2 │ {'embedding_dimension': 384.0} │ ollama │
├─────────────────┼─────────────────────────────────────┼─────────────────────────────────────┼───────────────────────────────────────────┼─────────────────┤ ├─────────────────┼─────────────────────────────────────┼─────────────────────────────────────┼───────────────────────────────────────────┼─────────────────┤
│ llm │ llama3.2:3b │ llama3.2:3b │ │ ollama │ │ llm │ meta-llama/Llama-3.2-3B-Instruct │ llama3.2:3b-instruct-fp16 │ │ ollama │
└─────────────────┴─────────────────────────────────────┴─────────────────────────────────────┴───────────────────────────────────────────┴─────────────────┘ └─────────────────┴─────────────────────────────────────┴─────────────────────────────────────┴───────────────────────────────────────────┴─────────────────┘
Total models: 2 Total models: 2

View file

@ -84,7 +84,7 @@ MODEL_ENTRIES = [
CoreModelId.llama_guard_3_1b.value, CoreModelId.llama_guard_3_1b.value,
), ),
ProviderModelEntry( ProviderModelEntry(
provider_model_id="all-minilm:latest", provider_model_id="all-minilm:l6-v2",
aliases=["all-minilm"], aliases=["all-minilm"],
model_type=ModelType.embedding, model_type=ModelType.embedding,
metadata={ metadata={

View file

@ -7,7 +7,7 @@ FROM --platform=linux/amd64 ollama/ollama:latest
RUN ollama serve & \ RUN ollama serve & \
sleep 5 && \ sleep 5 && \
ollama pull llama3.2:3b-instruct-fp16 && \ ollama pull llama3.2:3b-instruct-fp16 && \
ollama pull all-minilm:latest ollama pull all-minilm:l6-v2
# Set the entrypoint to start ollama serve # Set the entrypoint to start ollama serve
ENTRYPOINT ["ollama", "serve"] ENTRYPOINT ["ollama", "serve"]

View file

@ -105,7 +105,7 @@ models:
embedding_dimension: 384 embedding_dimension: 384
model_id: all-MiniLM-L6-v2 model_id: all-MiniLM-L6-v2
provider_id: custom_ollama provider_id: custom_ollama
provider_model_id: all-minilm:latest provider_model_id: all-minilm:l6-v2
model_type: embedding model_type: embedding
shields: [] shields: []
vector_dbs: [] vector_dbs: []