Updated the configuration templates to include the builtin preprocessors.

This commit is contained in:
ilya-kolchinsky 2025-03-07 16:08:14 +01:00
parent e895bb111c
commit 3f15349c9d
72 changed files with 632 additions and 25 deletions

View file

@ -29,4 +29,7 @@ distribution_spec:
- inline::code-interpreter
- inline::rag-runtime
- remote::model-context-protocol
preprocessing:
- inline::basic
- inline::simple_chunking
image_type: conda

View file

@ -7,6 +7,7 @@
from pathlib import Path
from llama_stack.apis.models.models import ModelType
from llama_stack.apis.preprocessing.preprocessors import PreprocessorInput
from llama_stack.distribution.datatypes import ModelInput, Provider, ToolGroupInput
from llama_stack.providers.inline.inference.meta_reference import (
MetaReferenceQuantizedInferenceConfig,
@ -35,6 +36,7 @@ def get_distribution_template() -> DistributionTemplate:
"inline::rag-runtime",
"remote::model-context-protocol",
],
"preprocessing": ["inline::basic", "inline::simple_chunking"],
}
default_tool_groups = [
ToolGroupInput(
@ -50,6 +52,16 @@ def get_distribution_template() -> DistributionTemplate:
provider_id="code-interpreter",
),
]
default_preprocessors = [
PreprocessorInput(
preprocessor_id="builtin::basic",
provider_id="basic",
),
PreprocessorInput(
preprocessor_id="builtin::chunking",
provider_id="simple_chunking",
),
]
name = "meta-reference-quantized-gpu"
inference_provider = Provider(
provider_id="meta-reference-inference",
@ -96,6 +108,7 @@ def get_distribution_template() -> DistributionTemplate:
},
default_models=[inference_model, embedding_model],
default_tool_groups=default_tool_groups,
default_preprocessors=default_preprocessors,
),
},
run_config_env_vars={

View file

@ -5,6 +5,7 @@ apis:
- datasetio
- eval
- inference
- preprocessing
- safety
- scoring
- telemetry
@ -92,6 +93,13 @@ providers:
- provider_id: model-context-protocol
provider_type: remote::model-context-protocol
config: {}
preprocessing:
- provider_id: basic
provider_type: inline::basic
config: {}
- provider_id: simple_chunking
provider_type: inline::simple_chunking
config: {}
metadata_store:
type: sqlite
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/meta-reference-quantized-gpu}/registry.db
@ -117,6 +125,10 @@ tool_groups:
provider_id: rag-runtime
- toolgroup_id: builtin::code_interpreter
provider_id: code-interpreter
preprocessors: []
preprocessors:
- preprocessor_id: builtin::basic
provider_id: basic
- preprocessor_id: builtin::chunking
provider_id: simple_chunking
server:
port: 8321