Update embedding dimension singular

This commit is contained in:
Ashwin Bharambe 2025-02-20 16:14:46 -08:00
parent 9436dd570d
commit 2608b6074f
9 changed files with 15 additions and 15 deletions

View file

@ -237,8 +237,8 @@ class FireworksInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProv
model = await self.model_store.get_model(model_id)
kwargs = {}
if model.metadata.get("embedding_dimensions"):
kwargs["dimensions"] = model.metadata.get("embedding_dimensions")
if model.metadata.get("embedding_dimension"):
kwargs["dimensions"] = model.metadata.get("embedding_dimension")
assert all(not content_has_media(content) for content in contents), (
"Fireworks does not support media for embeddings"
)

View file

@ -56,7 +56,7 @@ MODEL_ENTRIES = [
provider_model_id="nomic-ai/nomic-embed-text-v1.5",
model_type=ModelType.embedding,
metadata={
"embedding_dimensions": 768,
"embedding_dimension": 768,
"context_length": 8192,
},
),

View file

@ -88,7 +88,7 @@ model_entries = [
aliases=["all-minilm"],
model_type=ModelType.embedding,
metadata={
"embedding_dimensions": 384,
"embedding_dimension": 384,
"context_length": 512,
},
),
@ -96,7 +96,7 @@ model_entries = [
provider_model_id="nomic-embed-text",
model_type=ModelType.embedding,
metadata={
"embedding_dimensions": 768,
"embedding_dimension": 768,
"context_length": 8192,
},
),

View file

@ -52,7 +52,7 @@ MODEL_ENTRIES = [
provider_model_id="togethercomputer/m2-bert-80M-8k-retrieval",
model_type=ModelType.embedding,
metadata={
"embedding_dimensions": 768,
"embedding_dimension": 768,
"context_length": 8192,
},
),
@ -60,7 +60,7 @@ MODEL_ENTRIES = [
provider_model_id="togethercomputer/m2-bert-80M-32k-retrieval",
model_type=ModelType.embedding,
metadata={
"embedding_dimensions": 768,
"embedding_dimension": 768,
"context_length": 32768,
},
),

View file

@ -375,8 +375,8 @@ class VLLMInferenceAdapter(Inference, ModelsProtocolPrivate):
kwargs = {}
assert model.model_type == ModelType.embedding
assert model.metadata.get("embedding_dimensions")
kwargs["dimensions"] = model.metadata.get("embedding_dimensions")
assert model.metadata.get("embedding_dimension")
kwargs["dimensions"] = model.metadata.get("embedding_dimension")
assert all(not content_has_media(content) for content in contents), "VLLM does not support media for embeddings"
response = self.client.embeddings.create(
model=model.provider_resource_id,

View file

@ -150,7 +150,7 @@ models:
provider_model_id: accounts/fireworks/models/llama-guard-3-11b-vision
model_type: llm
- metadata:
embedding_dimensions: 768
embedding_dimension: 768
context_length: 8192
model_id: nomic-ai/nomic-embed-text-v1.5
provider_id: fireworks

View file

@ -144,7 +144,7 @@ models:
provider_model_id: accounts/fireworks/models/llama-guard-3-11b-vision
model_type: llm
- metadata:
embedding_dimensions: 768
embedding_dimension: 768
context_length: 8192
model_id: nomic-ai/nomic-embed-text-v1.5
provider_id: fireworks

View file

@ -145,14 +145,14 @@ models:
provider_model_id: meta-llama/Llama-Guard-3-11B-Vision-Turbo
model_type: llm
- metadata:
embedding_dimensions: 768
embedding_dimension: 768
context_length: 8192
model_id: togethercomputer/m2-bert-80M-8k-retrieval
provider_id: together
provider_model_id: togethercomputer/m2-bert-80M-8k-retrieval
model_type: embedding
- metadata:
embedding_dimensions: 768
embedding_dimension: 768
context_length: 32768
model_id: togethercomputer/m2-bert-80M-32k-retrieval
provider_id: together

View file

@ -139,14 +139,14 @@ models:
provider_model_id: meta-llama/Llama-Guard-3-11B-Vision-Turbo
model_type: llm
- metadata:
embedding_dimensions: 768
embedding_dimension: 768
context_length: 8192
model_id: togethercomputer/m2-bert-80M-8k-retrieval
provider_id: together
provider_model_id: togethercomputer/m2-bert-80M-8k-retrieval
model_type: embedding
- metadata:
embedding_dimensions: 768
embedding_dimension: 768
context_length: 32768
model_id: togethercomputer/m2-bert-80M-32k-retrieval
provider_id: together