mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-06 22:10:41 +00:00
fix(starter): Add missing faiss provider to build.yaml vector_io section (#2625)
The starter template build.yaml was missing the inline::faiss provider in the vector_io section, while it was properly configured in run.yaml and starter.py's vector_io_providers list. Fixes: #2624 Signed-off-by: Derek Higgins <derekh@redhat.com>
This commit is contained in:
parent
df6ce8befa
commit
4eae0cbfa4
2 changed files with 21 additions and 18 deletions
|
@ -29,6 +29,7 @@ distribution_spec:
|
||||||
- remote::passthrough
|
- remote::passthrough
|
||||||
- inline::sentence-transformers
|
- inline::sentence-transformers
|
||||||
vector_io:
|
vector_io:
|
||||||
|
- inline::faiss
|
||||||
- inline::sqlite-vec
|
- inline::sqlite-vec
|
||||||
- inline::milvus
|
- inline::milvus
|
||||||
- remote::chromadb
|
- remote::chromadb
|
||||||
|
|
|
@ -157,24 +157,7 @@ def get_remote_inference_providers() -> tuple[list[Provider], dict[str, list[Pro
|
||||||
|
|
||||||
def get_distribution_template() -> DistributionTemplate:
|
def get_distribution_template() -> DistributionTemplate:
|
||||||
remote_inference_providers, available_models = get_remote_inference_providers()
|
remote_inference_providers, available_models = get_remote_inference_providers()
|
||||||
providers = {
|
|
||||||
"inference": ([p.provider_type for p in remote_inference_providers] + ["inline::sentence-transformers"]),
|
|
||||||
"vector_io": ["inline::sqlite-vec", "inline::milvus", "remote::chromadb", "remote::pgvector"],
|
|
||||||
"files": ["inline::localfs"],
|
|
||||||
"safety": ["inline::llama-guard"],
|
|
||||||
"agents": ["inline::meta-reference"],
|
|
||||||
"telemetry": ["inline::meta-reference"],
|
|
||||||
"post_training": ["inline::huggingface"],
|
|
||||||
"eval": ["inline::meta-reference"],
|
|
||||||
"datasetio": ["remote::huggingface", "inline::localfs"],
|
|
||||||
"scoring": ["inline::basic", "inline::llm-as-judge", "inline::braintrust"],
|
|
||||||
"tool_runtime": [
|
|
||||||
"remote::brave-search",
|
|
||||||
"remote::tavily-search",
|
|
||||||
"inline::rag-runtime",
|
|
||||||
"remote::model-context-protocol",
|
|
||||||
],
|
|
||||||
}
|
|
||||||
name = "starter"
|
name = "starter"
|
||||||
|
|
||||||
vector_io_providers = [
|
vector_io_providers = [
|
||||||
|
@ -208,6 +191,25 @@ def get_distribution_template() -> DistributionTemplate:
|
||||||
),
|
),
|
||||||
),
|
),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
providers = {
|
||||||
|
"inference": ([p.provider_type for p in remote_inference_providers] + ["inline::sentence-transformers"]),
|
||||||
|
"vector_io": ([p.provider_type for p in vector_io_providers]),
|
||||||
|
"files": ["inline::localfs"],
|
||||||
|
"safety": ["inline::llama-guard"],
|
||||||
|
"agents": ["inline::meta-reference"],
|
||||||
|
"telemetry": ["inline::meta-reference"],
|
||||||
|
"post_training": ["inline::huggingface"],
|
||||||
|
"eval": ["inline::meta-reference"],
|
||||||
|
"datasetio": ["remote::huggingface", "inline::localfs"],
|
||||||
|
"scoring": ["inline::basic", "inline::llm-as-judge", "inline::braintrust"],
|
||||||
|
"tool_runtime": [
|
||||||
|
"remote::brave-search",
|
||||||
|
"remote::tavily-search",
|
||||||
|
"inline::rag-runtime",
|
||||||
|
"remote::model-context-protocol",
|
||||||
|
],
|
||||||
|
}
|
||||||
files_provider = Provider(
|
files_provider = Provider(
|
||||||
provider_id="meta-reference-files",
|
provider_id="meta-reference-files",
|
||||||
provider_type="inline::localfs",
|
provider_type="inline::localfs",
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue