mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
refactor: set proper name for embedding all-minilm:l6-v2 model
- we are using all-minilm:l6-v2 but the model we download from ollama is all-minilm:latest - even currently they are exactly the same model but if l12-v2 is updated, "latest" might not still valid. - the only change in this PR is pin the model from ollama Signed-off-by: Wen Zhou <wenzhou@redhat.com>
This commit is contained in:
parent
ea966565f6
commit
214b1fe1ae
4 changed files with 5 additions and 5 deletions
|
@ -180,9 +180,9 @@ Available Models
|
||||||
┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━┓
|
┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━┓
|
||||||
┃ model_type ┃ identifier ┃ provider_resource_id ┃ metadata ┃ provider_id ┃
|
┃ model_type ┃ identifier ┃ provider_resource_id ┃ metadata ┃ provider_id ┃
|
||||||
┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━┩
|
┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━┩
|
||||||
│ embedding │ all-MiniLM-L6-v2 │ all-minilm:latest │ {'embedding_dimension': 384.0} │ ollama │
|
│ embedding │ all-MiniLM-L6-v2 │ all-minilm:l6-v2 │ {'embedding_dimension': 384.0} │ ollama │
|
||||||
├─────────────────┼─────────────────────────────────────┼─────────────────────────────────────┼───────────────────────────────────────────┼─────────────────┤
|
├─────────────────┼─────────────────────────────────────┼─────────────────────────────────────┼───────────────────────────────────────────┼─────────────────┤
|
||||||
│ llm │ llama3.2:3b │ llama3.2:3b │ │ ollama │
|
│ llm │ meta-llama/Llama-3.2-3B-Instruct │ llama3.2:3b-instruct-fp16 │ │ ollama │
|
||||||
└─────────────────┴─────────────────────────────────────┴─────────────────────────────────────┴───────────────────────────────────────────┴─────────────────┘
|
└─────────────────┴─────────────────────────────────────┴─────────────────────────────────────┴───────────────────────────────────────────┴─────────────────┘
|
||||||
|
|
||||||
Total models: 2
|
Total models: 2
|
||||||
|
|
|
@ -84,7 +84,7 @@ MODEL_ENTRIES = [
|
||||||
CoreModelId.llama_guard_3_1b.value,
|
CoreModelId.llama_guard_3_1b.value,
|
||||||
),
|
),
|
||||||
ProviderModelEntry(
|
ProviderModelEntry(
|
||||||
provider_model_id="all-minilm:latest",
|
provider_model_id="all-minilm:l6-v2",
|
||||||
aliases=["all-minilm"],
|
aliases=["all-minilm"],
|
||||||
model_type=ModelType.embedding,
|
model_type=ModelType.embedding,
|
||||||
metadata={
|
metadata={
|
||||||
|
|
|
@ -7,7 +7,7 @@ FROM --platform=linux/amd64 ollama/ollama:latest
|
||||||
RUN ollama serve & \
|
RUN ollama serve & \
|
||||||
sleep 5 && \
|
sleep 5 && \
|
||||||
ollama pull llama3.2:3b-instruct-fp16 && \
|
ollama pull llama3.2:3b-instruct-fp16 && \
|
||||||
ollama pull all-minilm:latest
|
ollama pull all-minilm:l6-v2
|
||||||
|
|
||||||
# Set the entrypoint to start ollama serve
|
# Set the entrypoint to start ollama serve
|
||||||
ENTRYPOINT ["ollama", "serve"]
|
ENTRYPOINT ["ollama", "serve"]
|
||||||
|
|
|
@ -105,7 +105,7 @@ models:
|
||||||
embedding_dimension: 384
|
embedding_dimension: 384
|
||||||
model_id: all-MiniLM-L6-v2
|
model_id: all-MiniLM-L6-v2
|
||||||
provider_id: custom_ollama
|
provider_id: custom_ollama
|
||||||
provider_model_id: all-minilm:latest
|
provider_model_id: all-minilm:l6-v2
|
||||||
model_type: embedding
|
model_type: embedding
|
||||||
shields: []
|
shields: []
|
||||||
vector_dbs: []
|
vector_dbs: []
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue