Restrict the changes to the new preprocessing API only.

This commit is contained in:
ilya-kolchinsky 2025-04-03 12:19:08 +02:00
parent 2008cd7921
commit 863f87aa15
90 changed files with 104 additions and 1138 deletions

View file

@ -22,7 +22,4 @@ distribution_spec:
- inline::basic
tool_runtime:
- inline::rag-runtime
preprocessing:
- inline::basic
- inline::simple_chunking
image_type: conda

View file

@ -6,7 +6,6 @@
from pathlib import Path
from llama_stack.apis.preprocessors.preprocessors import PreprocessorInput
from llama_stack.distribution.datatypes import ModelInput, Provider, ShieldInput, ToolGroupInput
from llama_stack.providers.remote.inference.nvidia import NVIDIAConfig
from llama_stack.providers.remote.inference.nvidia.models import MODEL_ENTRIES
@ -26,7 +25,6 @@ def get_distribution_template() -> DistributionTemplate:
"datasetio": ["inline::localfs"],
"scoring": ["inline::basic"],
"tool_runtime": ["inline::rag-runtime"],
"preprocessing": ["inline::basic", "inline::simple_chunking"],
}
inference_provider = Provider(
@ -57,16 +55,6 @@ def get_distribution_template() -> DistributionTemplate:
provider_id="rag-runtime",
),
]
default_preprocessors = [
PreprocessorInput(
preprocessor_id="builtin::basic",
provider_id="basic",
),
PreprocessorInput(
preprocessor_id="builtin::chunking",
provider_id="simple_chunking",
),
]
default_models = get_model_registry(available_models)
return DistributionTemplate(
@ -84,7 +72,6 @@ def get_distribution_template() -> DistributionTemplate:
},
default_models=default_models,
default_tool_groups=default_tool_groups,
default_preprocessors=default_preprocessors,
),
"run-with-safety.yaml": RunConfigSettings(
provider_overrides={
@ -96,7 +83,6 @@ def get_distribution_template() -> DistributionTemplate:
default_models=[inference_model, safety_model],
default_shields=[ShieldInput(shield_id="${env.SAFETY_MODEL}", provider_id="nvidia")],
default_tool_groups=default_tool_groups,
default_preprocessors=default_preprocessors,
),
},
run_config_env_vars={

View file

@ -6,7 +6,6 @@ apis:
- eval
- inference
- post_training
- preprocessing
- safety
- scoring
- telemetry
@ -85,13 +84,6 @@ providers:
- provider_id: rag-runtime
provider_type: inline::rag-runtime
config: {}
preprocessing:
- provider_id: basic
provider_type: inline::basic
config: {}
- provider_id: simple_chunking
provider_type: inline::simple_chunking
config: {}
metadata_store:
type: sqlite
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/nvidia}/registry.db
@ -114,10 +106,6 @@ benchmarks: []
tool_groups:
- toolgroup_id: builtin::rag
provider_id: rag-runtime
preprocessors:
- preprocessor_id: builtin::basic
provider_id: basic
- preprocessor_id: builtin::chunking
provider_id: simple_chunking
preprocessors: []
server:
port: 8321

View file

@ -6,7 +6,6 @@ apis:
- eval
- inference
- post_training
- preprocessing
- safety
- scoring
- telemetry
@ -80,13 +79,6 @@ providers:
- provider_id: rag-runtime
provider_type: inline::rag-runtime
config: {}
preprocessing:
- provider_id: basic
provider_type: inline::basic
config: {}
- provider_id: simple_chunking
provider_type: inline::simple_chunking
config: {}
metadata_store:
type: sqlite
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/nvidia}/registry.db
@ -217,10 +209,6 @@ benchmarks: []
tool_groups:
- toolgroup_id: builtin::rag
provider_id: rag-runtime
preprocessors:
- preprocessor_id: builtin::basic
provider_id: basic
- preprocessor_id: builtin::chunking
provider_id: simple_chunking
preprocessors: []
server:
port: 8321