Update provider_type -> inline::llama-guard in templates, update run.yaml

This commit is contained in:
Ashwin Bharambe 2024-11-11 09:12:17 -08:00
parent 15ffceb533
commit 4971113f92
24 changed files with 121 additions and 98 deletions

View file

@ -11,6 +11,7 @@ from llama_stack.apis.shields import ShieldType
from llama_stack.distribution.datatypes import Api, Provider
from llama_stack.providers.inline.safety.llama_guard import LlamaGuardConfig
from llama_stack.providers.inline.safety.prompt_guard import PromptGuardConfig
from llama_stack.providers.remote.safety.bedrock import BedrockSafetyConfig
from llama_stack.providers.tests.resolver import resolve_impls_for_test_v2
@ -44,6 +45,22 @@ def safety_llama_guard(safety_model) -> ProviderFixture:
)
# TODO: this is not tested yet; we would need to configure the run_shield() test
# and parametrize it with the "prompt" for testing depending on the safety fixture
# we are using.
@pytest.fixture(scope="session")
def safety_prompt_guard() -> ProviderFixture:
return ProviderFixture(
providers=[
Provider(
provider_id="inline::prompt-guard",
provider_type="inline::prompt-guard",
config=PromptGuardConfig().model_dump(),
)
],
)
@pytest.fixture(scope="session")
def safety_bedrock() -> ProviderFixture:
return ProviderFixture(

View file

@ -3,7 +3,7 @@ distribution_spec:
description: Use Amazon Bedrock APIs.
providers:
inference: remote::bedrock
memory: meta-reference
safety: meta-reference
memory: inline::faiss
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -3,7 +3,7 @@ distribution_spec:
description: Use Databricks for running LLM inference
providers:
inference: remote::databricks
memory: meta-reference
safety: meta-reference
memory: inline::faiss
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -6,6 +6,6 @@ distribution_spec:
memory:
- meta-reference
- remote::weaviate
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -3,7 +3,7 @@ distribution_spec:
description: "Like local, but use Hugging Face Inference Endpoints for running LLM inference.\nSee https://hf.co/docs/api-endpoints."
providers:
inference: remote::hf::endpoint
memory: meta-reference
safety: meta-reference
memory: inline::faiss
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -3,7 +3,7 @@ distribution_spec:
description: "Like local, but use Hugging Face Inference API (serverless) for running LLM inference.\nSee https://hf.co/docs/api-inference."
providers:
inference: remote::hf::serverless
memory: meta-reference
safety: meta-reference
memory: inline::faiss
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -8,6 +8,6 @@ distribution_spec:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -8,6 +8,6 @@ distribution_spec:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -8,6 +8,6 @@ distribution_spec:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -7,6 +7,6 @@ distribution_spec:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -7,6 +7,6 @@ distribution_spec:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -7,6 +7,6 @@ distribution_spec:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference

View file

@ -6,6 +6,6 @@ distribution_spec:
memory:
- meta-reference
- remote::weaviate
safety: meta-reference
safety: inline::llama-guard
agents: meta-reference
telemetry: meta-reference