provider_id => provider_type, adapter_id => adapter_type

This commit is contained in:
Ashwin Bharambe 2024-10-02 14:05:59 -07:00
parent df68db644b
commit fe4aabd690
21 changed files with 83 additions and 85 deletions

View file

@ -4783,7 +4783,7 @@
"provider_config": { "provider_config": {
"type": "object", "type": "object",
"properties": { "properties": {
"provider_id": { "provider_type": {
"type": "string" "type": "string"
}, },
"config": { "config": {
@ -4814,7 +4814,7 @@
}, },
"additionalProperties": false, "additionalProperties": false,
"required": [ "required": [
"provider_id", "provider_type",
"config" "config"
] ]
} }
@ -4843,7 +4843,7 @@
"provider_config": { "provider_config": {
"type": "object", "type": "object",
"properties": { "properties": {
"provider_id": { "provider_type": {
"type": "string" "type": "string"
}, },
"config": { "config": {
@ -4874,7 +4874,7 @@
}, },
"additionalProperties": false, "additionalProperties": false,
"required": [ "required": [
"provider_id", "provider_type",
"config" "config"
] ]
} }
@ -4894,7 +4894,7 @@
"provider_config": { "provider_config": {
"type": "object", "type": "object",
"properties": { "properties": {
"provider_id": { "provider_type": {
"type": "string" "type": "string"
}, },
"config": { "config": {
@ -4925,7 +4925,7 @@
}, },
"additionalProperties": false, "additionalProperties": false,
"required": [ "required": [
"provider_id", "provider_type",
"config" "config"
] ]
} }

View file

@ -1117,10 +1117,10 @@ components:
- type: array - type: array
- type: object - type: object
type: object type: object
provider_id: provider_type:
type: string type: string
required: required:
- provider_id - provider_type
- config - config
type: object type: object
required: required:
@ -1362,10 +1362,10 @@ components:
- type: array - type: array
- type: object - type: object
type: object type: object
provider_id: provider_type:
type: string type: string
required: required:
- provider_id - provider_type
- config - config
type: object type: object
required: required:
@ -1916,10 +1916,10 @@ components:
- type: array - type: array
- type: object - type: object
type: object type: object
provider_id: provider_type:
type: string type: string
required: required:
- provider_id - provider_type
- config - config
type: object type: object
shield_type: shield_type:

View file

@ -18,7 +18,7 @@ from llama_stack.distribution.datatypes import GenericProviderConfig
class MemoryBankSpec(BaseModel): class MemoryBankSpec(BaseModel):
bank_type: MemoryBankType bank_type: MemoryBankType
provider_config: GenericProviderConfig = Field( provider_config: GenericProviderConfig = Field(
description="Provider config for the model, including provider_id, and corresponding config. ", description="Provider config for the model, including provider_type, and corresponding config. ",
) )

View file

@ -20,7 +20,7 @@ class ModelServingSpec(BaseModel):
description="All metadatas associated with llama model (defined in llama_models.models.sku_list).", description="All metadatas associated with llama model (defined in llama_models.models.sku_list).",
) )
provider_config: GenericProviderConfig = Field( provider_config: GenericProviderConfig = Field(
description="Provider config for the model, including provider_id, and corresponding config. ", description="Provider config for the model, including provider_type, and corresponding config. ",
) )

View file

@ -16,7 +16,7 @@ from llama_stack.distribution.datatypes import GenericProviderConfig
class ShieldSpec(BaseModel): class ShieldSpec(BaseModel):
shield_type: str shield_type: str
provider_config: GenericProviderConfig = Field( provider_config: GenericProviderConfig = Field(
description="Provider config for the model, including provider_id, and corresponding config. ", description="Provider config for the model, including provider_type, and corresponding config. ",
) )

View file

@ -47,11 +47,11 @@ class StackListProviders(Subcommand):
rows = [] rows = []
for spec in providers_for_api.values(): for spec in providers_for_api.values():
if spec.provider_id == "sample": if spec.provider_type == "sample":
continue continue
rows.append( rows.append(
[ [
spec.provider_id, spec.provider_type,
",".join(spec.pip_packages), ",".join(spec.pip_packages),
] ]
) )

View file

@ -109,7 +109,7 @@ def configure_api_providers(
routing_entries.append( routing_entries.append(
RoutableProviderConfig( RoutableProviderConfig(
routing_key=routing_key, routing_key=routing_key,
provider_id=p, provider_type=p,
config=cfg.dict(), config=cfg.dict(),
) )
) )
@ -120,7 +120,7 @@ def configure_api_providers(
routing_entries.append( routing_entries.append(
RoutableProviderConfig( RoutableProviderConfig(
routing_key=[s.value for s in MetaReferenceShieldType], routing_key=[s.value for s in MetaReferenceShieldType],
provider_id=p, provider_type=p,
config=cfg.dict(), config=cfg.dict(),
) )
) )
@ -133,7 +133,7 @@ def configure_api_providers(
routing_entries.append( routing_entries.append(
RoutableProviderConfig( RoutableProviderConfig(
routing_key=routing_key, routing_key=routing_key,
provider_id=p, provider_type=p,
config=cfg.dict(), config=cfg.dict(),
) )
) )
@ -153,7 +153,7 @@ def configure_api_providers(
routing_entries.append( routing_entries.append(
RoutableProviderConfig( RoutableProviderConfig(
routing_key=routing_key, routing_key=routing_key,
provider_id=p, provider_type=p,
config=cfg.dict(), config=cfg.dict(),
) )
) )
@ -164,7 +164,7 @@ def configure_api_providers(
) )
else: else:
config.api_providers[api_str] = GenericProviderConfig( config.api_providers[api_str] = GenericProviderConfig(
provider_id=p, provider_type=p,
config=cfg.dict(), config=cfg.dict(),
) )

View file

@ -71,7 +71,7 @@ Provider configurations for each of the APIs provided by this package.
E.g. The following is a ProviderRoutingEntry for models: E.g. The following is a ProviderRoutingEntry for models:
- routing_key: Meta-Llama3.1-8B-Instruct - routing_key: Meta-Llama3.1-8B-Instruct
provider_id: meta-reference provider_type: meta-reference
config: config:
model: Meta-Llama3.1-8B-Instruct model: Meta-Llama3.1-8B-Instruct
quantization: null quantization: null

View file

@ -51,7 +51,7 @@ def get_provider_registry() -> Dict[Api, Dict[str, ProviderSpec]]:
module = importlib.import_module(f"llama_stack.providers.registry.{name}") module = importlib.import_module(f"llama_stack.providers.registry.{name}")
ret[api] = { ret[api] = {
"remote": remote_provider_spec(api), "remote": remote_provider_spec(api),
**{a.provider_id: a for a in module.available_providers()}, **{a.provider_type: a for a in module.available_providers()},
} }
return ret return ret

View file

@ -18,10 +18,10 @@ class NeedsRequestProviderData:
spec = self.__provider_spec__ spec = self.__provider_spec__
assert spec, f"Provider spec not set on {self.__class__}" assert spec, f"Provider spec not set on {self.__class__}"
provider_id = spec.provider_id provider_type = spec.provider_type
validator_class = spec.provider_data_validator validator_class = spec.provider_data_validator
if not validator_class: if not validator_class:
raise ValueError(f"Provider {provider_id} does not have a validator") raise ValueError(f"Provider {provider_type} does not have a validator")
val = getattr(_THREAD_LOCAL, "provider_data_header_value", None) val = getattr(_THREAD_LOCAL, "provider_data_header_value", None)
if not val: if not val:

View file

@ -34,11 +34,11 @@ async def resolve_impls_with_routing(run_config: StackRunConfig) -> Dict[Api, An
if isinstance(config, PlaceholderProviderConfig): if isinstance(config, PlaceholderProviderConfig):
continue continue
if config.provider_id not in providers: if config.provider_type not in providers:
raise ValueError( raise ValueError(
f"Unknown provider `{config.provider_id}` is not available for API `{api}`" f"Provider `{config.provider_type}` is not available for API `{api}`"
) )
specs[api] = providers[config.provider_id] specs[api] = providers[config.provider_type]
configs[api] = config configs[api] = config
apis_to_serve = run_config.apis_to_serve or set( apis_to_serve = run_config.apis_to_serve or set(
@ -68,12 +68,12 @@ async def resolve_impls_with_routing(run_config: StackRunConfig) -> Dict[Api, An
inner_specs = [] inner_specs = []
inner_deps = [] inner_deps = []
for rt_entry in routing_table: for rt_entry in routing_table:
if rt_entry.provider_id not in providers: if rt_entry.provider_type not in providers:
raise ValueError( raise ValueError(
f"Unknown provider `{rt_entry.provider_id}` is not available for API `{api}`" f"Provider `{rt_entry.provider_type}` is not available for API `{api}`"
) )
inner_specs.append(providers[rt_entry.provider_id]) inner_specs.append(providers[rt_entry.provider_type])
inner_deps.extend(providers[rt_entry.provider_id].api_dependencies) inner_deps.extend(providers[rt_entry.provider_type].api_dependencies)
specs[source_api] = RoutingTableProviderSpec( specs[source_api] = RoutingTableProviderSpec(
api=source_api, api=source_api,
@ -94,7 +94,7 @@ async def resolve_impls_with_routing(run_config: StackRunConfig) -> Dict[Api, An
sorted_specs = topological_sort(specs.values()) sorted_specs = topological_sort(specs.values())
print(f"Resolved {len(sorted_specs)} providers in topological order") print(f"Resolved {len(sorted_specs)} providers in topological order")
for spec in sorted_specs: for spec in sorted_specs:
print(f" {spec.api}: {spec.provider_id}") print(f" {spec.api}: {spec.provider_type}")
print("") print("")
impls = {} impls = {}
for spec in sorted_specs: for spec in sorted_specs:

View file

@ -18,7 +18,7 @@ api_providers:
providers: providers:
- meta-reference - meta-reference
agents: agents:
provider_id: meta-reference provider_type: meta-reference
config: config:
persistence_store: persistence_store:
namespace: null namespace: null
@ -28,22 +28,22 @@ api_providers:
providers: providers:
- meta-reference - meta-reference
telemetry: telemetry:
provider_id: meta-reference provider_type: meta-reference
config: {} config: {}
routing_table: routing_table:
inference: inference:
- provider_id: remote::ollama - provider_type: remote::ollama
config: config:
host: localhost host: localhost
port: 6000 port: 6000
routing_key: Meta-Llama3.1-8B-Instruct routing_key: Meta-Llama3.1-8B-Instruct
safety: safety:
- provider_id: meta-reference - provider_type: meta-reference
config: config:
llama_guard_shield: null llama_guard_shield: null
prompt_guard_shield: null prompt_guard_shield: null
routing_key: ["llama_guard", "code_scanner_guard", "injection_shield", "jailbreak_shield"] routing_key: ["llama_guard", "code_scanner_guard", "injection_shield", "jailbreak_shield"]
memory: memory:
- provider_id: meta-reference - provider_type: meta-reference
config: {} config: {}
routing_key: vector routing_key: vector

View file

@ -18,7 +18,7 @@ api_providers:
providers: providers:
- meta-reference - meta-reference
agents: agents:
provider_id: meta-reference provider_type: meta-reference
config: config:
persistence_store: persistence_store:
namespace: null namespace: null
@ -28,11 +28,11 @@ api_providers:
providers: providers:
- meta-reference - meta-reference
telemetry: telemetry:
provider_id: meta-reference provider_type: meta-reference
config: {} config: {}
routing_table: routing_table:
inference: inference:
- provider_id: meta-reference - provider_type: meta-reference
config: config:
model: Llama3.1-8B-Instruct model: Llama3.1-8B-Instruct
quantization: null quantization: null
@ -41,12 +41,12 @@ routing_table:
max_batch_size: 1 max_batch_size: 1
routing_key: Llama3.1-8B-Instruct routing_key: Llama3.1-8B-Instruct
safety: safety:
- provider_id: meta-reference - provider_type: meta-reference
config: config:
llama_guard_shield: null llama_guard_shield: null
prompt_guard_shield: null prompt_guard_shield: null
routing_key: ["llama_guard", "code_scanner_guard", "injection_shield", "jailbreak_shield"] routing_key: ["llama_guard", "code_scanner_guard", "injection_shield", "jailbreak_shield"]
memory: memory:
- provider_id: meta-reference - provider_type: meta-reference
config: {} config: {}
routing_key: vector routing_key: vector

View file

@ -46,11 +46,11 @@ async def instantiate_provider(
assert isinstance(provider_config, List) assert isinstance(provider_config, List)
routing_table = provider_config routing_table = provider_config
inner_specs = {x.provider_id: x for x in provider_spec.inner_specs} inner_specs = {x.provider_type: x for x in provider_spec.inner_specs}
inner_impls = [] inner_impls = []
for routing_entry in routing_table: for routing_entry in routing_table:
impl = await instantiate_provider( impl = await instantiate_provider(
inner_specs[routing_entry.provider_id], inner_specs[routing_entry.provider_type],
deps, deps,
routing_entry, routing_entry,
) )

View file

@ -28,7 +28,7 @@ class Api(Enum):
@json_schema_type @json_schema_type
class ProviderSpec(BaseModel): class ProviderSpec(BaseModel):
api: Api api: Api
provider_id: str provider_type: str
config_class: str = Field( config_class: str = Field(
..., ...,
description="Fully-qualified classname of the config for this provider", description="Fully-qualified classname of the config for this provider",
@ -56,7 +56,7 @@ class RoutableProvider(Protocol):
class GenericProviderConfig(BaseModel): class GenericProviderConfig(BaseModel):
provider_id: str provider_type: str
config: Dict[str, Any] config: Dict[str, Any]
@ -76,7 +76,7 @@ class RoutableProviderConfig(GenericProviderConfig):
# Example: /inference, /safety # Example: /inference, /safety
@json_schema_type @json_schema_type
class AutoRoutedProviderSpec(ProviderSpec): class AutoRoutedProviderSpec(ProviderSpec):
provider_id: str = "router" provider_type: str = "router"
config_class: str = "" config_class: str = ""
docker_image: Optional[str] = None docker_image: Optional[str] = None
@ -101,7 +101,7 @@ class AutoRoutedProviderSpec(ProviderSpec):
# Example: /models, /shields # Example: /models, /shields
@json_schema_type @json_schema_type
class RoutingTableProviderSpec(ProviderSpec): class RoutingTableProviderSpec(ProviderSpec):
provider_id: str = "routing_table" provider_type: str = "routing_table"
config_class: str = "" config_class: str = ""
docker_image: Optional[str] = None docker_image: Optional[str] = None
@ -119,7 +119,7 @@ class RoutingTableProviderSpec(ProviderSpec):
@json_schema_type @json_schema_type
class AdapterSpec(BaseModel): class AdapterSpec(BaseModel):
adapter_id: str = Field( adapter_type: str = Field(
..., ...,
description="Unique identifier for this adapter", description="Unique identifier for this adapter",
) )
@ -179,8 +179,8 @@ class RemoteProviderConfig(BaseModel):
return f"http://{self.host}:{self.port}" return f"http://{self.host}:{self.port}"
def remote_provider_id(adapter_id: str) -> str: def remote_provider_type(adapter_type: str) -> str:
return f"remote::{adapter_id}" return f"remote::{adapter_type}"
@json_schema_type @json_schema_type
@ -226,8 +226,8 @@ def remote_provider_spec(
if adapter and adapter.config_class if adapter and adapter.config_class
else "llama_stack.distribution.datatypes.RemoteProviderConfig" else "llama_stack.distribution.datatypes.RemoteProviderConfig"
) )
provider_id = remote_provider_id(adapter.adapter_id) if adapter else "remote" provider_type = remote_provider_type(adapter.adapter_type) if adapter else "remote"
return RemoteProviderSpec( return RemoteProviderSpec(
api=api, provider_id=provider_id, config_class=config_class, adapter=adapter api=api, provider_type=provider_type, config_class=config_class, adapter=adapter
) )

View file

@ -14,7 +14,7 @@ def available_providers() -> List[ProviderSpec]:
return [ return [
InlineProviderSpec( InlineProviderSpec(
api=Api.agents, api=Api.agents,
provider_id="meta-reference", provider_type="meta-reference",
pip_packages=[ pip_packages=[
"matplotlib", "matplotlib",
"pillow", "pillow",
@ -33,7 +33,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.agents, api=Api.agents,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="sample", adapter_type="sample",
pip_packages=[], pip_packages=[],
module="llama_stack.providers.adapters.agents.sample", module="llama_stack.providers.adapters.agents.sample",
config_class="llama_stack.providers.adapters.agents.sample.SampleConfig", config_class="llama_stack.providers.adapters.agents.sample.SampleConfig",

View file

@ -13,7 +13,7 @@ def available_providers() -> List[ProviderSpec]:
return [ return [
InlineProviderSpec( InlineProviderSpec(
api=Api.inference, api=Api.inference,
provider_id="meta-reference", provider_type="meta-reference",
pip_packages=[ pip_packages=[
"accelerate", "accelerate",
"blobfile", "blobfile",
@ -30,7 +30,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="sample", adapter_type="sample",
pip_packages=[], pip_packages=[],
module="llama_stack.providers.adapters.inference.sample", module="llama_stack.providers.adapters.inference.sample",
config_class="llama_stack.providers.adapters.inference.sample.SampleConfig", config_class="llama_stack.providers.adapters.inference.sample.SampleConfig",
@ -39,7 +39,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="ollama", adapter_type="ollama",
pip_packages=["ollama"], pip_packages=["ollama"],
module="llama_stack.providers.adapters.inference.ollama", module="llama_stack.providers.adapters.inference.ollama",
), ),
@ -47,7 +47,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="tgi", adapter_type="tgi",
pip_packages=["huggingface_hub", "aiohttp"], pip_packages=["huggingface_hub", "aiohttp"],
module="llama_stack.providers.adapters.inference.tgi", module="llama_stack.providers.adapters.inference.tgi",
config_class="llama_stack.providers.adapters.inference.tgi.TGIImplConfig", config_class="llama_stack.providers.adapters.inference.tgi.TGIImplConfig",
@ -56,7 +56,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="hf::serverless", adapter_type="hf::serverless",
pip_packages=["huggingface_hub", "aiohttp"], pip_packages=["huggingface_hub", "aiohttp"],
module="llama_stack.providers.adapters.inference.tgi", module="llama_stack.providers.adapters.inference.tgi",
config_class="llama_stack.providers.adapters.inference.tgi.InferenceAPIImplConfig", config_class="llama_stack.providers.adapters.inference.tgi.InferenceAPIImplConfig",
@ -65,7 +65,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="hf::endpoint", adapter_type="hf::endpoint",
pip_packages=["huggingface_hub", "aiohttp"], pip_packages=["huggingface_hub", "aiohttp"],
module="llama_stack.providers.adapters.inference.tgi", module="llama_stack.providers.adapters.inference.tgi",
config_class="llama_stack.providers.adapters.inference.tgi.InferenceEndpointImplConfig", config_class="llama_stack.providers.adapters.inference.tgi.InferenceEndpointImplConfig",
@ -74,7 +74,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="fireworks", adapter_type="fireworks",
pip_packages=[ pip_packages=[
"fireworks-ai", "fireworks-ai",
], ],
@ -85,7 +85,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="together", adapter_type="together",
pip_packages=[ pip_packages=[
"together", "together",
], ],
@ -97,10 +97,8 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.inference, api=Api.inference,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="bedrock", adapter_type="bedrock",
pip_packages=[ pip_packages=["boto3"],
"boto3"
],
module="llama_stack.providers.adapters.inference.bedrock", module="llama_stack.providers.adapters.inference.bedrock",
config_class="llama_stack.providers.adapters.inference.bedrock.BedrockConfig", config_class="llama_stack.providers.adapters.inference.bedrock.BedrockConfig",
), ),

View file

@ -34,7 +34,7 @@ def available_providers() -> List[ProviderSpec]:
return [ return [
InlineProviderSpec( InlineProviderSpec(
api=Api.memory, api=Api.memory,
provider_id="meta-reference", provider_type="meta-reference",
pip_packages=EMBEDDING_DEPS + ["faiss-cpu"], pip_packages=EMBEDDING_DEPS + ["faiss-cpu"],
module="llama_stack.providers.impls.meta_reference.memory", module="llama_stack.providers.impls.meta_reference.memory",
config_class="llama_stack.providers.impls.meta_reference.memory.FaissImplConfig", config_class="llama_stack.providers.impls.meta_reference.memory.FaissImplConfig",
@ -42,7 +42,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
Api.memory, Api.memory,
AdapterSpec( AdapterSpec(
adapter_id="chromadb", adapter_type="chromadb",
pip_packages=EMBEDDING_DEPS + ["chromadb-client"], pip_packages=EMBEDDING_DEPS + ["chromadb-client"],
module="llama_stack.providers.adapters.memory.chroma", module="llama_stack.providers.adapters.memory.chroma",
), ),
@ -50,7 +50,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
Api.memory, Api.memory,
AdapterSpec( AdapterSpec(
adapter_id="pgvector", adapter_type="pgvector",
pip_packages=EMBEDDING_DEPS + ["psycopg2-binary"], pip_packages=EMBEDDING_DEPS + ["psycopg2-binary"],
module="llama_stack.providers.adapters.memory.pgvector", module="llama_stack.providers.adapters.memory.pgvector",
config_class="llama_stack.providers.adapters.memory.pgvector.PGVectorConfig", config_class="llama_stack.providers.adapters.memory.pgvector.PGVectorConfig",
@ -59,7 +59,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.memory, api=Api.memory,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="sample", adapter_type="sample",
pip_packages=[], pip_packages=[],
module="llama_stack.providers.adapters.memory.sample", module="llama_stack.providers.adapters.memory.sample",
config_class="llama_stack.providers.adapters.memory.sample.SampleConfig", config_class="llama_stack.providers.adapters.memory.sample.SampleConfig",

View file

@ -19,7 +19,7 @@ def available_providers() -> List[ProviderSpec]:
return [ return [
InlineProviderSpec( InlineProviderSpec(
api=Api.safety, api=Api.safety,
provider_id="meta-reference", provider_type="meta-reference",
pip_packages=[ pip_packages=[
"codeshield", "codeshield",
"transformers", "transformers",
@ -34,7 +34,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.safety, api=Api.safety,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="sample", adapter_type="sample",
pip_packages=[], pip_packages=[],
module="llama_stack.providers.adapters.safety.sample", module="llama_stack.providers.adapters.safety.sample",
config_class="llama_stack.providers.adapters.safety.sample.SampleConfig", config_class="llama_stack.providers.adapters.safety.sample.SampleConfig",
@ -43,7 +43,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.safety, api=Api.safety,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="bedrock", adapter_type="bedrock",
pip_packages=["boto3"], pip_packages=["boto3"],
module="llama_stack.providers.adapters.safety.bedrock", module="llama_stack.providers.adapters.safety.bedrock",
config_class="llama_stack.providers.adapters.safety.bedrock.BedrockSafetyConfig", config_class="llama_stack.providers.adapters.safety.bedrock.BedrockSafetyConfig",
@ -52,7 +52,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.safety, api=Api.safety,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="together", adapter_type="together",
pip_packages=[ pip_packages=[
"together", "together",
], ],

View file

@ -13,7 +13,7 @@ def available_providers() -> List[ProviderSpec]:
return [ return [
InlineProviderSpec( InlineProviderSpec(
api=Api.telemetry, api=Api.telemetry,
provider_id="meta-reference", provider_type="meta-reference",
pip_packages=[], pip_packages=[],
module="llama_stack.providers.impls.meta_reference.telemetry", module="llama_stack.providers.impls.meta_reference.telemetry",
config_class="llama_stack.providers.impls.meta_reference.telemetry.ConsoleConfig", config_class="llama_stack.providers.impls.meta_reference.telemetry.ConsoleConfig",
@ -21,7 +21,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.telemetry, api=Api.telemetry,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="sample", adapter_type="sample",
pip_packages=[], pip_packages=[],
module="llama_stack.providers.adapters.telemetry.sample", module="llama_stack.providers.adapters.telemetry.sample",
config_class="llama_stack.providers.adapters.telemetry.sample.SampleConfig", config_class="llama_stack.providers.adapters.telemetry.sample.SampleConfig",
@ -30,7 +30,7 @@ def available_providers() -> List[ProviderSpec]:
remote_provider_spec( remote_provider_spec(
api=Api.telemetry, api=Api.telemetry,
adapter=AdapterSpec( adapter=AdapterSpec(
adapter_id="opentelemetry-jaeger", adapter_type="opentelemetry-jaeger",
pip_packages=[ pip_packages=[
"opentelemetry-api", "opentelemetry-api",
"opentelemetry-sdk", "opentelemetry-sdk",

View file

@ -18,7 +18,7 @@ api_providers:
providers: providers:
- meta-reference - meta-reference
agents: agents:
provider_id: meta-reference provider_type: meta-reference
config: config:
persistence_store: persistence_store:
namespace: null namespace: null
@ -28,11 +28,11 @@ api_providers:
providers: providers:
- meta-reference - meta-reference
telemetry: telemetry:
provider_id: meta-reference provider_type: meta-reference
config: {} config: {}
routing_table: routing_table:
inference: inference:
- provider_id: meta-reference - provider_type: meta-reference
config: config:
model: Meta-Llama3.1-8B-Instruct model: Meta-Llama3.1-8B-Instruct
quantization: null quantization: null
@ -41,7 +41,7 @@ routing_table:
max_batch_size: 1 max_batch_size: 1
routing_key: Meta-Llama3.1-8B-Instruct routing_key: Meta-Llama3.1-8B-Instruct
safety: safety:
- provider_id: meta-reference - provider_type: meta-reference
config: config:
llama_guard_shield: llama_guard_shield:
model: Llama-Guard-3-1B model: Llama-Guard-3-1B
@ -52,6 +52,6 @@ routing_table:
model: Prompt-Guard-86M model: Prompt-Guard-86M
routing_key: ["llama_guard", "code_scanner_guard", "injection_shield", "jailbreak_shield"] routing_key: ["llama_guard", "code_scanner_guard", "injection_shield", "jailbreak_shield"]
memory: memory:
- provider_id: meta-reference - provider_type: meta-reference
config: {} config: {}
routing_key: vector routing_key: vector