refactor: Add ProviderContext for a flexible storage directory

- Introduce ProviderContext class to decouple provider storage paths from absolute paths
- Add storage_dir attribute to StackRunConfig to accept CLI options
- Implement storage directory resolution with prioritized fallbacks:
  1. CLI option (--state-directory)
  2. Environment variable (LLAMA_STACK_STATE_DIR)
  3. Default distribution directory
- Standardize provider signatures to follow context, config, deps pattern
- Update provider implementations to use the new context-based approach
- Add comprehensive tests to verify state directory resolution
This commit is contained in:
Roland Huß 2025-05-12 11:44:21 +02:00
parent dd07c7a5b5
commit e6c9aebe47
41 changed files with 242 additions and 81 deletions

View file

@ -6,12 +6,14 @@
from typing import Any
from llama_stack.providers.datatypes import ProviderContext
from .config import CodeScannerConfig
async def get_provider_impl(config: CodeScannerConfig, deps: dict[str, Any]):
async def get_provider_impl(context: ProviderContext, config: CodeScannerConfig, deps: dict[str, Any]):
from .code_scanner import MetaReferenceCodeScannerSafetyImpl
impl = MetaReferenceCodeScannerSafetyImpl(config, deps)
impl = MetaReferenceCodeScannerSafetyImpl(context, config, deps)
await impl.initialize()
return impl

View file

@ -15,6 +15,7 @@ from llama_stack.apis.safety import (
ViolationLevel,
)
from llama_stack.apis.shields import Shield
from llama_stack.providers.datatypes import ProviderContext
from llama_stack.providers.utils.inference.prompt_adapter import (
interleaved_content_as_str,
)
@ -30,8 +31,10 @@ ALLOWED_CODE_SCANNER_MODEL_IDS = [
class MetaReferenceCodeScannerSafetyImpl(Safety):
def __init__(self, config: CodeScannerConfig, deps) -> None:
def __init__(self, context: ProviderContext, config: CodeScannerConfig, deps) -> None:
self.context = context
self.config = config
self.deps = deps
async def initialize(self) -> None:
pass

View file

@ -6,14 +6,16 @@
from typing import Any
from llama_stack.providers.datatypes import ProviderContext
from .config import LlamaGuardConfig
async def get_provider_impl(config: LlamaGuardConfig, deps: dict[str, Any]):
async def get_provider_impl(context: ProviderContext, config: LlamaGuardConfig, deps: dict[str, Any]):
from .llama_guard import LlamaGuardSafetyImpl
assert isinstance(config, LlamaGuardConfig), f"Unexpected config type: {type(config)}"
impl = LlamaGuardSafetyImpl(config, deps)
impl = LlamaGuardSafetyImpl(context, config, deps)
await impl.initialize()
return impl

View file

@ -24,7 +24,7 @@ from llama_stack.apis.shields import Shield
from llama_stack.distribution.datatypes import Api
from llama_stack.models.llama.datatypes import Role
from llama_stack.models.llama.sku_types import CoreModelId
from llama_stack.providers.datatypes import ShieldsProtocolPrivate
from llama_stack.providers.datatypes import ProviderContext, ShieldsProtocolPrivate
from llama_stack.providers.utils.inference.prompt_adapter import (
interleaved_content_as_str,
)
@ -130,7 +130,8 @@ PROMPT_TEMPLATE = Template(f"{PROMPT_TASK}{SAFETY_CATEGORIES}{PROMPT_CONVERSATIO
class LlamaGuardSafetyImpl(Safety, ShieldsProtocolPrivate):
def __init__(self, config: LlamaGuardConfig, deps) -> None:
def __init__(self, context: ProviderContext, config: LlamaGuardConfig, deps) -> None:
self.context = context
self.config = config
self.inference_api = deps[Api.inference]

View file

@ -6,12 +6,14 @@
from typing import Any
from llama_stack.providers.datatypes import ProviderContext
from .config import PromptGuardConfig
async def get_provider_impl(config: PromptGuardConfig, deps: dict[str, Any]):
async def get_provider_impl(context: ProviderContext, config: PromptGuardConfig, deps: dict[str, Any]):
from .prompt_guard import PromptGuardSafetyImpl
impl = PromptGuardSafetyImpl(config, deps)
impl = PromptGuardSafetyImpl(context, config, deps)
await impl.initialize()
return impl

View file

@ -19,7 +19,7 @@ from llama_stack.apis.safety import (
)
from llama_stack.apis.shields import Shield
from llama_stack.distribution.utils.model_utils import model_local_dir
from llama_stack.providers.datatypes import ShieldsProtocolPrivate
from llama_stack.providers.datatypes import ProviderContext, ShieldsProtocolPrivate
from llama_stack.providers.utils.inference.prompt_adapter import (
interleaved_content_as_str,
)
@ -32,8 +32,10 @@ PROMPT_GUARD_MODEL = "Prompt-Guard-86M"
class PromptGuardSafetyImpl(Safety, ShieldsProtocolPrivate):
def __init__(self, config: PromptGuardConfig, _deps) -> None:
def __init__(self, context: ProviderContext, config: PromptGuardConfig, _deps) -> None:
self.context = context
self.config = config
self.deps = _deps
async def initialize(self) -> None:
model_dir = model_local_dir(PROMPT_GUARD_MODEL)