[memory refactor][5/n] Migrate all vector_io providers (#835)

See https://github.com/meta-llama/llama-stack/issues/827 for the broader
design.

This PR finishes off all the stragglers and migrates everything to the
new naming.
This commit is contained in:
Ashwin Bharambe 2025-01-22 10:17:59 -08:00 committed by GitHub
parent 63f37f9b7c
commit c9e5578151
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
78 changed files with 504 additions and 623 deletions

View file

@ -4,7 +4,7 @@ distribution_spec:
providers:
inference:
- remote::nvidia
memory:
vector_io:
- inline::faiss
safety:
- inline::llama-guard

View file

@ -17,7 +17,7 @@ from llama_stack.templates.template import DistributionTemplate, RunConfigSettin
def get_distribution_template() -> DistributionTemplate:
providers = {
"inference": ["remote::nvidia"],
"memory": ["inline::faiss"],
"vector_io": ["inline::faiss"],
"safety": ["inline::llama-guard"],
"agents": ["inline::meta-reference"],
"telemetry": ["inline::meta-reference"],

View file

@ -5,11 +5,11 @@ apis:
- datasetio
- eval
- inference
- memory
- safety
- scoring
- telemetry
- tool_runtime
- vector_io
providers:
inference:
- provider_id: nvidia
@ -17,7 +17,7 @@ providers:
config:
url: https://integrate.api.nvidia.com
api_key: ${env.NVIDIA_API_KEY}
memory:
vector_io:
- provider_id: faiss
provider_type: inline::faiss
config:
@ -136,7 +136,7 @@ models:
provider_model_id: meta/llama-3.2-90b-vision-instruct
model_type: llm
shields: []
memory_banks: []
vector_dbs: []
datasets: []
scoring_fns: []
eval_tasks: []