This commit is contained in:
Xi Yan 2025-03-18 21:55:06 -07:00
parent b3060ce8a1
commit 583de3d80c
3 changed files with 6 additions and 11 deletions

View file

@ -106,7 +106,7 @@ class CommonRoutingTableImpl(RoutingTable):
await self.dist_registry.register(obj) await self.dist_registry.register(obj)
# Register all objects from providers # Register all objects from providers
for pid, p in self.impls_by_provider_id.items(): for _pid, p in self.impls_by_provider_id.items():
api = get_impl_api(p) api = get_impl_api(p)
if api == Api.inference: if api == Api.inference:
p.model_store = self p.model_store = self

View file

@ -13,7 +13,6 @@ from llama_stack.apis.benchmarks import Benchmark
from llama_stack.apis.datasetio import DatasetIO from llama_stack.apis.datasetio import DatasetIO
from llama_stack.apis.datasets import Datasets from llama_stack.apis.datasets import Datasets
from llama_stack.apis.inference import Inference, SystemMessage, UserMessage from llama_stack.apis.inference import Inference, SystemMessage, UserMessage
from llama_stack.apis.scoring import Scoring
from llama_stack.providers.datatypes import BenchmarksProtocolPrivate from llama_stack.providers.datatypes import BenchmarksProtocolPrivate
from llama_stack.providers.inline.agents.meta_reference.agent_instance import ( from llama_stack.providers.inline.agents.meta_reference.agent_instance import (
MEMORY_QUERY_TOOL, MEMORY_QUERY_TOOL,
@ -37,14 +36,14 @@ class MetaReferenceEvalImpl(
config: MetaReferenceEvalConfig, config: MetaReferenceEvalConfig,
datasetio_api: DatasetIO, datasetio_api: DatasetIO,
datasets_api: Datasets, datasets_api: Datasets,
scoring_api: Scoring,
inference_api: Inference, inference_api: Inference,
agents_api: Agents, agents_api: Agents,
) -> None: ) -> None:
self.config = config self.config = config
self.datasetio_api = datasetio_api self.datasetio_api = datasetio_api
self.datasets_api = datasets_api self.datasets_api = datasets_api
self.scoring_api = scoring_api # TODO(xiyan): this implementation will be refactored
self.scoring_api = None
self.inference_api = inference_api self.inference_api = inference_api
self.agents_api = agents_api self.agents_api = agents_api

View file

@ -30,14 +30,12 @@ from llama_stack.providers.remote.vector_io.pgvector.config import (
from llama_stack.providers.utils.inference.model_registry import ProviderModelEntry from llama_stack.providers.utils.inference.model_registry import ProviderModelEntry
from llama_stack.templates.template import ( from llama_stack.templates.template import (
DistributionTemplate, DistributionTemplate,
get_model_registry,
RunConfigSettings, RunConfigSettings,
get_model_registry,
) )
def get_inference_providers() -> ( def get_inference_providers() -> Tuple[List[Provider], Dict[str, List[ProviderModelEntry]]]:
Tuple[List[Provider], Dict[str, List[ProviderModelEntry]]]
):
# in this template, we allow each API key to be optional # in this template, we allow each API key to be optional
providers = [ providers = [
( (
@ -118,9 +116,7 @@ def get_distribution_template() -> DistributionTemplate:
Provider( Provider(
provider_id="sqlite-vec", provider_id="sqlite-vec",
provider_type="inline::sqlite-vec", provider_type="inline::sqlite-vec",
config=SQLiteVectorIOConfig.sample_run_config( config=SQLiteVectorIOConfig.sample_run_config(f"~/.llama/distributions/{name}"),
f"~/.llama/distributions/{name}"
),
), ),
Provider( Provider(
provider_id="${env.ENABLE_CHROMADB+chromadb}", provider_id="${env.ENABLE_CHROMADB+chromadb}",