diff --git a/llama_stack/templates/together/run-with-safety.yaml b/llama_stack/templates/together/run-with-safety.yaml index 52ee36e34..0c72d4986 100644 --- a/llama_stack/templates/together/run-with-safety.yaml +++ b/llama_stack/templates/together/run-with-safety.yaml @@ -150,6 +150,36 @@ models: provider_id: together provider_model_id: togethercomputer/m2-bert-80M-8k-retrieval model_type: embedding +- metadata: + embedding_dimension: 768 + model_id: togethercomputer/m2-bert-80M-2k-retrieval + provider_id: together + provider_model_id: togethercomputer/m2-bert-80M-2k-retrieval + model_type: embedding +- metadata: + embedding_dimension: 768 + model_id: togethercomputer/m2-bert-80M-32k-retrieval + provider_id: together + provider_model_id: togethercomputer/m2-bert-80M-32k-retrieval + model_type: embedding +- metadata: + embedding_dimension: 1024 + model_id: WhereIsAI/UAE-Large-V1 + provider_id: together + provider_model_id: WhereIsAI/UAE-Large-V1 + model_type: embedding +- metadata: + embedding_dimension: 1024 + model_id: BAAI/bge-large-en-v1.5 + provider_id: together + provider_model_id: BAAI/bge-large-en-v1.5 + model_type: embedding +- metadata: + embedding_dimension: 768 + model_id: BAAI/bge-base-en-v1.5 + provider_id: together + provider_model_id: BAAI/bge-base-en-v1.5 + model_type: embedding shields: - shield_id: meta-llama/Llama-Guard-3-8B provider_id: llama-guard diff --git a/llama_stack/templates/together/run.yaml b/llama_stack/templates/together/run.yaml index d9fc5c5fc..88aee0da3 100644 --- a/llama_stack/templates/together/run.yaml +++ b/llama_stack/templates/together/run.yaml @@ -144,6 +144,36 @@ models: provider_id: together provider_model_id: togethercomputer/m2-bert-80M-8k-retrieval model_type: embedding +- metadata: + embedding_dimension: 768 + model_id: togethercomputer/m2-bert-80M-2k-retrieval + provider_id: together + provider_model_id: togethercomputer/m2-bert-80M-2k-retrieval + model_type: embedding +- metadata: + embedding_dimension: 768 + model_id: togethercomputer/m2-bert-80M-32k-retrieval + provider_id: together + provider_model_id: togethercomputer/m2-bert-80M-32k-retrieval + model_type: embedding +- metadata: + embedding_dimension: 1024 + model_id: WhereIsAI/UAE-Large-V1 + provider_id: together + provider_model_id: WhereIsAI/UAE-Large-V1 + model_type: embedding +- metadata: + embedding_dimension: 1024 + model_id: BAAI/bge-large-en-v1.5 + provider_id: together + provider_model_id: BAAI/bge-large-en-v1.5 + model_type: embedding +- metadata: + embedding_dimension: 768 + model_id: BAAI/bge-base-en-v1.5 + provider_id: together + provider_model_id: BAAI/bge-base-en-v1.5 + model_type: embedding shields: - shield_id: meta-llama/Llama-Guard-3-8B vector_dbs: [] diff --git a/llama_stack/templates/together/together.py b/llama_stack/templates/together/together.py index 4c571db05..3ba443e39 100644 --- a/llama_stack/templates/together/together.py +++ b/llama_stack/templates/together/together.py @@ -81,13 +81,50 @@ def get_distribution_template() -> DistributionTemplate: provider_id="code-interpreter", ), ] - embedding_model = ModelInput( - model_id="togethercomputer/m2-bert-80M-8k-retrieval", - provider_id="together", - model_type=ModelType.embedding, - provider_model_id="togethercomputer/m2-bert-80M-8k-retrieval", - metadata={"embedding_dimension": 768}, - ) + embedding_models = [ + ModelInput( + model_id="togethercomputer/m2-bert-80M-8k-retrieval", + provider_id="together", + model_type=ModelType.embedding, + provider_model_id="togethercomputer/m2-bert-80M-8k-retrieval", + metadata={"embedding_dimension": 768}, + ), + ModelInput( + model_id="togethercomputer/m2-bert-80M-2k-retrieval", + provider_id="together", + model_type=ModelType.embedding, + provider_model_id="togethercomputer/m2-bert-80M-2k-retrieval", + metadata={"embedding_dimension": 768}, + ), + ModelInput( + model_id="togethercomputer/m2-bert-80M-32k-retrieval", + provider_id="together", + model_type=ModelType.embedding, + provider_model_id="togethercomputer/m2-bert-80M-32k-retrieval", + metadata={"embedding_dimension": 768}, + ), + ModelInput( + model_id="WhereIsAI/UAE-Large-V1", + provider_id="together", + model_type=ModelType.embedding, + provider_model_id="WhereIsAI/UAE-Large-V1", + metadata={"embedding_dimension": 1024}, + ), + ModelInput( + model_id="BAAI/bge-large-en-v1.5", + provider_id="together", + model_type=ModelType.embedding, + provider_model_id="BAAI/bge-large-en-v1.5", + metadata={"embedding_dimension": 1024}, + ), + ModelInput( + model_id="BAAI/bge-base-en-v1.5", + provider_id="together", + model_type=ModelType.embedding, + provider_model_id="BAAI/bge-base-en-v1.5", + metadata={"embedding_dimension": 768}, + ), + ] return DistributionTemplate( name=name, @@ -103,7 +140,7 @@ def get_distribution_template() -> DistributionTemplate: "inference": [inference_provider, embedding_provider], "vector_io": [vector_io_provider], }, - default_models=default_models + [embedding_model], + default_models=default_models + embedding_models, default_tool_groups=default_tool_groups, default_shields=[ShieldInput(shield_id="meta-llama/Llama-Guard-3-8B")], ), @@ -134,7 +171,7 @@ def get_distribution_template() -> DistributionTemplate: }, default_models=[ *default_models, - embedding_model, + *embedding_models, ], default_shields=[ ShieldInput(