From 6562a409801292d80a2f236f34c3fd43f3934d5c Mon Sep 17 00:00:00 2001 From: Kaushik Date: Tue, 11 Feb 2025 15:10:52 -0800 Subject: [PATCH] Revert "add fiddlecube to distros, update docs" This reverts commit 3978df54b627402e6d8f47d4d94fd0eb2b4899d1. --- docs/source/distributions/remote_hosted_distro/nvidia.md | 2 +- docs/source/distributions/self_hosted_distro/bedrock.md | 2 +- docs/source/distributions/self_hosted_distro/cerebras.md | 2 +- docs/source/distributions/self_hosted_distro/dell.md | 2 +- .../distributions/self_hosted_distro/meta-reference-gpu.md | 2 +- .../self_hosted_distro/meta-reference-quantized-gpu.md | 2 +- docs/source/distributions/self_hosted_distro/ollama.md | 2 +- docs/source/distributions/self_hosted_distro/remote-vllm.md | 2 +- docs/source/distributions/self_hosted_distro/sambanova.md | 2 +- docs/source/distributions/self_hosted_distro/tgi.md | 2 +- docs/source/distributions/self_hosted_distro/together.md | 2 +- docs/source/index.md | 1 - llama_stack/templates/bedrock/bedrock.py | 2 +- llama_stack/templates/bedrock/build.yaml | 1 - llama_stack/templates/bedrock/run.yaml | 3 --- llama_stack/templates/cerebras/build.yaml | 1 - llama_stack/templates/cerebras/cerebras.py | 2 +- llama_stack/templates/cerebras/run.yaml | 3 --- llama_stack/templates/dell/build.yaml | 1 - llama_stack/templates/dell/dell.py | 2 +- llama_stack/templates/dell/run-with-safety.yaml | 3 --- llama_stack/templates/dell/run.yaml | 3 --- llama_stack/templates/fireworks/build.yaml | 1 - llama_stack/templates/fireworks/fireworks.py | 2 +- llama_stack/templates/fireworks/run.yaml | 4 ++-- llama_stack/templates/hf-endpoint/build.yaml | 1 - llama_stack/templates/hf-endpoint/hf_endpoint.py | 2 +- llama_stack/templates/hf-endpoint/run-with-safety.yaml | 3 --- llama_stack/templates/hf-endpoint/run.yaml | 3 --- llama_stack/templates/hf-serverless/build.yaml | 1 - llama_stack/templates/hf-serverless/hf_serverless.py | 2 +- llama_stack/templates/hf-serverless/run-with-safety.yaml | 3 --- llama_stack/templates/hf-serverless/run.yaml | 3 --- llama_stack/templates/meta-reference-gpu/build.yaml | 1 - llama_stack/templates/meta-reference-gpu/meta_reference.py | 2 +- llama_stack/templates/meta-reference-gpu/run-with-safety.yaml | 3 --- llama_stack/templates/meta-reference-gpu/run.yaml | 3 --- llama_stack/templates/meta-reference-quantized-gpu/build.yaml | 1 - .../templates/meta-reference-quantized-gpu/meta_reference.py | 2 +- llama_stack/templates/meta-reference-quantized-gpu/run.yaml | 3 --- llama_stack/templates/nvidia/build.yaml | 1 - llama_stack/templates/nvidia/nvidia.py | 2 +- llama_stack/templates/nvidia/run.yaml | 3 --- llama_stack/templates/ollama/build.yaml | 1 - llama_stack/templates/ollama/ollama.py | 2 +- llama_stack/templates/ollama/run.yaml | 3 --- llama_stack/templates/remote-vllm/build.yaml | 1 - llama_stack/templates/remote-vllm/run-with-safety.yaml | 3 --- llama_stack/templates/remote-vllm/run.yaml | 3 --- llama_stack/templates/remote-vllm/vllm.py | 2 +- llama_stack/templates/sambanova/build.yaml | 1 - llama_stack/templates/sambanova/run.yaml | 3 --- llama_stack/templates/sambanova/sambanova.py | 2 +- llama_stack/templates/tgi/build.yaml | 1 - llama_stack/templates/tgi/run-with-safety.yaml | 3 --- llama_stack/templates/tgi/run.yaml | 3 --- llama_stack/templates/tgi/tgi.py | 2 +- llama_stack/templates/together/build.yaml | 1 - llama_stack/templates/together/run.yaml | 3 --- llama_stack/templates/together/together.py | 2 +- llama_stack/templates/vllm-gpu/build.yaml | 1 - llama_stack/templates/vllm-gpu/run.yaml | 3 --- llama_stack/templates/vllm-gpu/vllm.py | 2 +- 63 files changed, 28 insertions(+), 104 deletions(-) diff --git a/docs/source/distributions/remote_hosted_distro/nvidia.md b/docs/source/distributions/remote_hosted_distro/nvidia.md index 70e334041..f352f737e 100644 --- a/docs/source/distributions/remote_hosted_distro/nvidia.md +++ b/docs/source/distributions/remote_hosted_distro/nvidia.md @@ -9,7 +9,7 @@ The `llamastack/distribution-nvidia` distribution consists of the following prov | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::nvidia` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/distributions/self_hosted_distro/bedrock.md b/docs/source/distributions/self_hosted_distro/bedrock.md index 3f6dc3a70..64c9f8c19 100644 --- a/docs/source/distributions/self_hosted_distro/bedrock.md +++ b/docs/source/distributions/self_hosted_distro/bedrock.md @@ -16,7 +16,7 @@ The `llamastack/distribution-bedrock` distribution consists of the following pro | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::bedrock` | -| safety | `remote::bedrock`, `remote::fiddlecube` | +| safety | `remote::bedrock` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/distributions/self_hosted_distro/cerebras.md b/docs/source/distributions/self_hosted_distro/cerebras.md index c1aa608c3..a0c9eb263 100644 --- a/docs/source/distributions/self_hosted_distro/cerebras.md +++ b/docs/source/distributions/self_hosted_distro/cerebras.md @@ -9,7 +9,7 @@ The `llamastack/distribution-cerebras` distribution consists of the following pr | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::cerebras` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | diff --git a/docs/source/distributions/self_hosted_distro/dell.md b/docs/source/distributions/self_hosted_distro/dell.md index 74ad5afcc..aef3ecf58 100644 --- a/docs/source/distributions/self_hosted_distro/dell.md +++ b/docs/source/distributions/self_hosted_distro/dell.md @@ -20,7 +20,7 @@ The `llamastack/distribution-dell` distribution consists of the following provid | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::tgi` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | diff --git a/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md b/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md index d449c4a5c..b183757db 100644 --- a/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md +++ b/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md @@ -19,7 +19,7 @@ The `llamastack/distribution-meta-reference-gpu` distribution consists of the fo | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `inline::meta-reference` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md b/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md index 6a37ad061..9aeb7a88b 100644 --- a/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md +++ b/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md @@ -19,7 +19,7 @@ The `llamastack/distribution-meta-reference-quantized-gpu` distribution consists | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `inline::meta-reference-quantized` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/distributions/self_hosted_distro/ollama.md b/docs/source/distributions/self_hosted_distro/ollama.md index ff1cf9379..a3a45f9a8 100644 --- a/docs/source/distributions/self_hosted_distro/ollama.md +++ b/docs/source/distributions/self_hosted_distro/ollama.md @@ -19,7 +19,7 @@ The `llamastack/distribution-ollama` distribution consists of the following prov | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::ollama` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | diff --git a/docs/source/distributions/self_hosted_distro/remote-vllm.md b/docs/source/distributions/self_hosted_distro/remote-vllm.md index f109bd03e..6c3bbd1d0 100644 --- a/docs/source/distributions/self_hosted_distro/remote-vllm.md +++ b/docs/source/distributions/self_hosted_distro/remote-vllm.md @@ -18,7 +18,7 @@ The `llamastack/distribution-remote-vllm` distribution consists of the following | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::vllm` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/distributions/self_hosted_distro/sambanova.md b/docs/source/distributions/self_hosted_distro/sambanova.md index 5bc5c9c98..e6ac616be 100644 --- a/docs/source/distributions/self_hosted_distro/sambanova.md +++ b/docs/source/distributions/self_hosted_distro/sambanova.md @@ -17,7 +17,7 @@ The `llamastack/distribution-sambanova` distribution consists of the following p |-----|-------------| | agents | `inline::meta-reference` | | inference | `remote::sambanova` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/tgi.md b/docs/source/distributions/self_hosted_distro/tgi.md index f77aa1352..f4eecf2cd 100644 --- a/docs/source/distributions/self_hosted_distro/tgi.md +++ b/docs/source/distributions/self_hosted_distro/tgi.md @@ -20,7 +20,7 @@ The `llamastack/distribution-tgi` distribution consists of the following provide | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::tgi` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/distributions/self_hosted_distro/together.md b/docs/source/distributions/self_hosted_distro/together.md index da331f5a6..8e36c1eb0 100644 --- a/docs/source/distributions/self_hosted_distro/together.md +++ b/docs/source/distributions/self_hosted_distro/together.md @@ -19,7 +19,7 @@ The `llamastack/distribution-together` distribution consists of the following pr | datasetio | `remote::huggingface`, `inline::localfs` | | eval | `inline::meta-reference` | | inference | `remote::together` | -| safety | `inline::llama-guard`, `remote::fiddlecube` | +| safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | | tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | diff --git a/docs/source/index.md b/docs/source/index.md index c431e1441..2834f5641 100644 --- a/docs/source/index.md +++ b/docs/source/index.md @@ -78,7 +78,6 @@ A number of "adapters" are available for some popular Inference and Vector Store | Prompt Guard | Single Node | | Code Scanner | Single Node | | AWS Bedrock | Hosted | -| FiddleCube | Hosted | ```{toctree} diff --git a/llama_stack/templates/bedrock/bedrock.py b/llama_stack/templates/bedrock/bedrock.py index 62003e383..0c8259285 100644 --- a/llama_stack/templates/bedrock/bedrock.py +++ b/llama_stack/templates/bedrock/bedrock.py @@ -19,7 +19,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::bedrock"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["remote::bedrock", "remote::fiddlecube"], + "safety": ["remote::bedrock"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/bedrock/build.yaml b/llama_stack/templates/bedrock/build.yaml index f121a25e9..6c07b0478 100644 --- a/llama_stack/templates/bedrock/build.yaml +++ b/llama_stack/templates/bedrock/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - remote::bedrock - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/bedrock/run.yaml b/llama_stack/templates/bedrock/run.yaml index 916238463..be6c9a928 100644 --- a/llama_stack/templates/bedrock/run.yaml +++ b/llama_stack/templates/bedrock/run.yaml @@ -27,9 +27,6 @@ providers: - provider_id: bedrock provider_type: remote::bedrock config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/cerebras/build.yaml b/llama_stack/templates/cerebras/build.yaml index 27eb97761..9d5ab1a52 100644 --- a/llama_stack/templates/cerebras/build.yaml +++ b/llama_stack/templates/cerebras/build.yaml @@ -6,7 +6,6 @@ distribution_spec: - remote::cerebras safety: - inline::llama-guard - - remote::fiddlecube vector_io: - inline::faiss - remote::chromadb diff --git a/llama_stack/templates/cerebras/cerebras.py b/llama_stack/templates/cerebras/cerebras.py index 0e491f300..2dfae04f8 100644 --- a/llama_stack/templates/cerebras/cerebras.py +++ b/llama_stack/templates/cerebras/cerebras.py @@ -22,7 +22,7 @@ from llama_stack.templates.template import DistributionTemplate, RunConfigSettin def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::cerebras"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], "agents": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/cerebras/run.yaml b/llama_stack/templates/cerebras/run.yaml index fda928406..05d3f4525 100644 --- a/llama_stack/templates/cerebras/run.yaml +++ b/llama_stack/templates/cerebras/run.yaml @@ -24,9 +24,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} vector_io: - provider_id: faiss provider_type: inline::faiss diff --git a/llama_stack/templates/dell/build.yaml b/llama_stack/templates/dell/build.yaml index be1fe9a6b..e2edb9386 100644 --- a/llama_stack/templates/dell/build.yaml +++ b/llama_stack/templates/dell/build.yaml @@ -11,7 +11,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/dell/dell.py b/llama_stack/templates/dell/dell.py index 2c7cdbb9a..5781da7f4 100644 --- a/llama_stack/templates/dell/dell.py +++ b/llama_stack/templates/dell/dell.py @@ -23,7 +23,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::tgi"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/dell/run-with-safety.yaml b/llama_stack/templates/dell/run-with-safety.yaml index 21bf0cc77..04c5957d4 100644 --- a/llama_stack/templates/dell/run-with-safety.yaml +++ b/llama_stack/templates/dell/run-with-safety.yaml @@ -32,9 +32,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/dell/run.yaml b/llama_stack/templates/dell/run.yaml index 585baba34..706444eb1 100644 --- a/llama_stack/templates/dell/run.yaml +++ b/llama_stack/templates/dell/run.yaml @@ -28,9 +28,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/fireworks/build.yaml b/llama_stack/templates/fireworks/build.yaml index adbfc4da0..cdd60ec2a 100644 --- a/llama_stack/templates/fireworks/build.yaml +++ b/llama_stack/templates/fireworks/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/fireworks/fireworks.py b/llama_stack/templates/fireworks/fireworks.py index 785d3e2de..ec350010b 100644 --- a/llama_stack/templates/fireworks/fireworks.py +++ b/llama_stack/templates/fireworks/fireworks.py @@ -28,7 +28,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::fireworks"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/fireworks/run.yaml b/llama_stack/templates/fireworks/run.yaml index 5f13e07e7..ccf67dcbb 100644 --- a/llama_stack/templates/fireworks/run.yaml +++ b/llama_stack/templates/fireworks/run.yaml @@ -29,8 +29,8 @@ providers: namespace: null db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/fireworks}/faiss_store.db safety: - - provider_id: fiddlecube - provider_type: remote::fiddlecube + - provider_id: llama-guard + provider_type: inline::llama-guard config: {} agents: - provider_id: meta-reference diff --git a/llama_stack/templates/hf-endpoint/build.yaml b/llama_stack/templates/hf-endpoint/build.yaml index bca47e9a4..c2eaaa05b 100644 --- a/llama_stack/templates/hf-endpoint/build.yaml +++ b/llama_stack/templates/hf-endpoint/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/hf-endpoint/hf_endpoint.py b/llama_stack/templates/hf-endpoint/hf_endpoint.py index ea2d43ad5..4533fd95b 100644 --- a/llama_stack/templates/hf-endpoint/hf_endpoint.py +++ b/llama_stack/templates/hf-endpoint/hf_endpoint.py @@ -23,7 +23,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::hf::endpoint"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/hf-endpoint/run-with-safety.yaml b/llama_stack/templates/hf-endpoint/run-with-safety.yaml index 5555f9ffb..f520a2fda 100644 --- a/llama_stack/templates/hf-endpoint/run-with-safety.yaml +++ b/llama_stack/templates/hf-endpoint/run-with-safety.yaml @@ -37,9 +37,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/hf-endpoint/run.yaml b/llama_stack/templates/hf-endpoint/run.yaml index d3962fcbd..708cb1bcc 100644 --- a/llama_stack/templates/hf-endpoint/run.yaml +++ b/llama_stack/templates/hf-endpoint/run.yaml @@ -32,9 +32,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/hf-serverless/build.yaml b/llama_stack/templates/hf-serverless/build.yaml index c037ca018..f9303cfab 100644 --- a/llama_stack/templates/hf-serverless/build.yaml +++ b/llama_stack/templates/hf-serverless/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/hf-serverless/hf_serverless.py b/llama_stack/templates/hf-serverless/hf_serverless.py index 14f59d56e..8438de7a5 100644 --- a/llama_stack/templates/hf-serverless/hf_serverless.py +++ b/llama_stack/templates/hf-serverless/hf_serverless.py @@ -23,7 +23,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::hf::serverless"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/hf-serverless/run-with-safety.yaml b/llama_stack/templates/hf-serverless/run-with-safety.yaml index 66d209a1a..7f0abf5be 100644 --- a/llama_stack/templates/hf-serverless/run-with-safety.yaml +++ b/llama_stack/templates/hf-serverless/run-with-safety.yaml @@ -37,9 +37,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/hf-serverless/run.yaml b/llama_stack/templates/hf-serverless/run.yaml index 497b08da3..c0b7a4c60 100644 --- a/llama_stack/templates/hf-serverless/run.yaml +++ b/llama_stack/templates/hf-serverless/run.yaml @@ -32,9 +32,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/meta-reference-gpu/build.yaml b/llama_stack/templates/meta-reference-gpu/build.yaml index 15dcc8200..b9130fc7d 100644 --- a/llama_stack/templates/meta-reference-gpu/build.yaml +++ b/llama_stack/templates/meta-reference-gpu/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/meta-reference-gpu/meta_reference.py b/llama_stack/templates/meta-reference-gpu/meta_reference.py index b1f240287..a3f82b0c8 100644 --- a/llama_stack/templates/meta-reference-gpu/meta_reference.py +++ b/llama_stack/templates/meta-reference-gpu/meta_reference.py @@ -27,7 +27,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["inline::meta-reference"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml b/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml index 6c2b97b47..c5286fc6b 100644 --- a/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml +++ b/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml @@ -39,9 +39,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/meta-reference-gpu/run.yaml b/llama_stack/templates/meta-reference-gpu/run.yaml index 63912f098..310585f23 100644 --- a/llama_stack/templates/meta-reference-gpu/run.yaml +++ b/llama_stack/templates/meta-reference-gpu/run.yaml @@ -33,9 +33,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/meta-reference-quantized-gpu/build.yaml b/llama_stack/templates/meta-reference-quantized-gpu/build.yaml index e939e6ca6..7bbcfe5f2 100644 --- a/llama_stack/templates/meta-reference-quantized-gpu/build.yaml +++ b/llama_stack/templates/meta-reference-quantized-gpu/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py b/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py index 71fcc46f1..8c2a6ec9f 100644 --- a/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py +++ b/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py @@ -22,7 +22,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["inline::meta-reference-quantized"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/meta-reference-quantized-gpu/run.yaml b/llama_stack/templates/meta-reference-quantized-gpu/run.yaml index 93fbbe721..d43cf3917 100644 --- a/llama_stack/templates/meta-reference-quantized-gpu/run.yaml +++ b/llama_stack/templates/meta-reference-quantized-gpu/run.yaml @@ -35,9 +35,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/nvidia/build.yaml b/llama_stack/templates/nvidia/build.yaml index f36f7ee35..e9748721a 100644 --- a/llama_stack/templates/nvidia/build.yaml +++ b/llama_stack/templates/nvidia/build.yaml @@ -8,7 +8,6 @@ distribution_spec: - inline::faiss safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/nvidia/nvidia.py b/llama_stack/templates/nvidia/nvidia.py index 2d82fcfe7..d24c9ed48 100644 --- a/llama_stack/templates/nvidia/nvidia.py +++ b/llama_stack/templates/nvidia/nvidia.py @@ -18,7 +18,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::nvidia"], "vector_io": ["inline::faiss"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/nvidia/run.yaml b/llama_stack/templates/nvidia/run.yaml index cce4b7868..c8ae362f5 100644 --- a/llama_stack/templates/nvidia/run.yaml +++ b/llama_stack/templates/nvidia/run.yaml @@ -29,9 +29,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/ollama/build.yaml b/llama_stack/templates/ollama/build.yaml index 3305ddd97..0fee6808c 100644 --- a/llama_stack/templates/ollama/build.yaml +++ b/llama_stack/templates/ollama/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/ollama/ollama.py b/llama_stack/templates/ollama/ollama.py index c30da5549..d14cb3aad 100644 --- a/llama_stack/templates/ollama/ollama.py +++ b/llama_stack/templates/ollama/ollama.py @@ -25,7 +25,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::ollama"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/ollama/run.yaml b/llama_stack/templates/ollama/run.yaml index 0c887b903..485223675 100644 --- a/llama_stack/templates/ollama/run.yaml +++ b/llama_stack/templates/ollama/run.yaml @@ -31,9 +31,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/remote-vllm/build.yaml b/llama_stack/templates/remote-vllm/build.yaml index 6e04368a6..74d9f32d9 100644 --- a/llama_stack/templates/remote-vllm/build.yaml +++ b/llama_stack/templates/remote-vllm/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference eval: diff --git a/llama_stack/templates/remote-vllm/run-with-safety.yaml b/llama_stack/templates/remote-vllm/run-with-safety.yaml index 19d956585..1fe998a1f 100644 --- a/llama_stack/templates/remote-vllm/run-with-safety.yaml +++ b/llama_stack/templates/remote-vllm/run-with-safety.yaml @@ -39,9 +39,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/remote-vllm/run.yaml b/llama_stack/templates/remote-vllm/run.yaml index bc05cfa45..9d3db8a31 100644 --- a/llama_stack/templates/remote-vllm/run.yaml +++ b/llama_stack/templates/remote-vllm/run.yaml @@ -33,9 +33,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/remote-vllm/vllm.py b/llama_stack/templates/remote-vllm/vllm.py index 66b602a84..6c835ef86 100644 --- a/llama_stack/templates/remote-vllm/vllm.py +++ b/llama_stack/templates/remote-vllm/vllm.py @@ -25,7 +25,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::vllm"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "eval": ["inline::meta-reference"], "datasetio": ["remote::huggingface", "inline::localfs"], diff --git a/llama_stack/templates/sambanova/build.yaml b/llama_stack/templates/sambanova/build.yaml index ed2bdb4ea..ca5ffe618 100644 --- a/llama_stack/templates/sambanova/build.yaml +++ b/llama_stack/templates/sambanova/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/sambanova/run.yaml b/llama_stack/templates/sambanova/run.yaml index 7c02e4c9c..39b0f3c4e 100644 --- a/llama_stack/templates/sambanova/run.yaml +++ b/llama_stack/templates/sambanova/run.yaml @@ -32,9 +32,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/sambanova/sambanova.py b/llama_stack/templates/sambanova/sambanova.py index 66e3ea0ae..70b54b010 100644 --- a/llama_stack/templates/sambanova/sambanova.py +++ b/llama_stack/templates/sambanova/sambanova.py @@ -24,7 +24,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::sambanova"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "tool_runtime": [ diff --git a/llama_stack/templates/tgi/build.yaml b/llama_stack/templates/tgi/build.yaml index a61627eb6..8bc628158 100644 --- a/llama_stack/templates/tgi/build.yaml +++ b/llama_stack/templates/tgi/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/tgi/run-with-safety.yaml b/llama_stack/templates/tgi/run-with-safety.yaml index e4015d1f6..ed6c9ef6f 100644 --- a/llama_stack/templates/tgi/run-with-safety.yaml +++ b/llama_stack/templates/tgi/run-with-safety.yaml @@ -32,9 +32,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/tgi/run.yaml b/llama_stack/templates/tgi/run.yaml index cece8d83e..8bf76f37b 100644 --- a/llama_stack/templates/tgi/run.yaml +++ b/llama_stack/templates/tgi/run.yaml @@ -31,9 +31,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/tgi/tgi.py b/llama_stack/templates/tgi/tgi.py index cbfa36ab7..e49c98d72 100644 --- a/llama_stack/templates/tgi/tgi.py +++ b/llama_stack/templates/tgi/tgi.py @@ -25,7 +25,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::tgi"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/together/build.yaml b/llama_stack/templates/together/build.yaml index ce5bec920..90ee5bcee 100644 --- a/llama_stack/templates/together/build.yaml +++ b/llama_stack/templates/together/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/together/run.yaml b/llama_stack/templates/together/run.yaml index 8ce2eae8b..920003759 100644 --- a/llama_stack/templates/together/run.yaml +++ b/llama_stack/templates/together/run.yaml @@ -32,9 +32,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/together/together.py b/llama_stack/templates/together/together.py index 1f724fae3..b7ac130ed 100644 --- a/llama_stack/templates/together/together.py +++ b/llama_stack/templates/together/together.py @@ -28,7 +28,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["remote::together"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"], diff --git a/llama_stack/templates/vllm-gpu/build.yaml b/llama_stack/templates/vllm-gpu/build.yaml index fb38fed74..d24046613 100644 --- a/llama_stack/templates/vllm-gpu/build.yaml +++ b/llama_stack/templates/vllm-gpu/build.yaml @@ -10,7 +10,6 @@ distribution_spec: - remote::pgvector safety: - inline::llama-guard - - remote::fiddlecube agents: - inline::meta-reference telemetry: diff --git a/llama_stack/templates/vllm-gpu/run.yaml b/llama_stack/templates/vllm-gpu/run.yaml index bf56c3e49..41a545e1a 100644 --- a/llama_stack/templates/vllm-gpu/run.yaml +++ b/llama_stack/templates/vllm-gpu/run.yaml @@ -35,9 +35,6 @@ providers: - provider_id: llama-guard provider_type: inline::llama-guard config: {} - - provider_id: fiddlecube - provider_type: remote::fiddlecube - config: {} agents: - provider_id: meta-reference provider_type: inline::meta-reference diff --git a/llama_stack/templates/vllm-gpu/vllm.py b/llama_stack/templates/vllm-gpu/vllm.py index fff73924b..54ebd2d41 100644 --- a/llama_stack/templates/vllm-gpu/vllm.py +++ b/llama_stack/templates/vllm-gpu/vllm.py @@ -22,7 +22,7 @@ def get_distribution_template() -> DistributionTemplate: providers = { "inference": ["inline::vllm"], "vector_io": ["inline::faiss", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard", "remote::fiddlecube"], + "safety": ["inline::llama-guard"], "agents": ["inline::meta-reference"], "telemetry": ["inline::meta-reference"], "eval": ["inline::meta-reference"],