mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-16 06:53:47 +00:00
Update provider_type -> inline::llama-guard in templates, update run.yaml
This commit is contained in:
parent
15ffceb533
commit
4971113f92
24 changed files with 121 additions and 98 deletions
|
@ -3,7 +3,7 @@ distribution_spec:
|
|||
description: "Like local, but use Hugging Face Inference API (serverless) for running LLM inference.\nSee https://hf.co/docs/api-inference."
|
||||
providers:
|
||||
inference: remote::hf::serverless
|
||||
memory: meta-reference
|
||||
safety: meta-reference
|
||||
memory: inline::faiss
|
||||
safety: inline::llama-guard
|
||||
agents: meta-reference
|
||||
telemetry: meta-reference
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue