From ec1a308dac5cf9c96befc102090ec1f6de8c17c2 Mon Sep 17 00:00:00 2001 From: Charlie Doern Date: Mon, 1 Dec 2025 11:49:41 -0500 Subject: [PATCH] fix: remove all build.yaml files Signed-off-by: Charlie Doern --- .../distributions/ci-tests/build.yaml | 55 ------------------ src/llama_stack/distributions/dell/build.yaml | 30 ---------- .../meta-reference-gpu/build.yaml | 29 ---------- .../distributions/nvidia/build.yaml | 26 --------- src/llama_stack/distributions/oci/build.yaml | 32 ----------- .../distributions/open-benchmark/build.yaml | 33 ----------- .../distributions/starter-gpu/build.yaml | 56 ------------------- .../distributions/starter/build.yaml | 56 ------------------- .../distributions/watsonx/build.yaml | 30 ---------- tests/external/build.yaml | 10 ---- tests/external/ramalama-stack/build.yaml | 10 ---- 11 files changed, 367 deletions(-) delete mode 100644 src/llama_stack/distributions/ci-tests/build.yaml delete mode 100644 src/llama_stack/distributions/dell/build.yaml delete mode 100644 src/llama_stack/distributions/meta-reference-gpu/build.yaml delete mode 100644 src/llama_stack/distributions/nvidia/build.yaml delete mode 100644 src/llama_stack/distributions/oci/build.yaml delete mode 100644 src/llama_stack/distributions/open-benchmark/build.yaml delete mode 100644 src/llama_stack/distributions/starter-gpu/build.yaml delete mode 100644 src/llama_stack/distributions/starter/build.yaml delete mode 100644 src/llama_stack/distributions/watsonx/build.yaml delete mode 100644 tests/external/build.yaml delete mode 100644 tests/external/ramalama-stack/build.yaml diff --git a/src/llama_stack/distributions/ci-tests/build.yaml b/src/llama_stack/distributions/ci-tests/build.yaml deleted file mode 100644 index a4a9b6354..000000000 --- a/src/llama_stack/distributions/ci-tests/build.yaml +++ /dev/null @@ -1,55 +0,0 @@ -version: 2 -distribution_spec: - description: CI tests for Llama Stack - providers: - inference: - - provider_type: remote::cerebras - - provider_type: remote::ollama - - provider_type: remote::vllm - - provider_type: remote::tgi - - provider_type: remote::fireworks - - provider_type: remote::together - - provider_type: remote::bedrock - - provider_type: remote::nvidia - - provider_type: remote::openai - - provider_type: remote::anthropic - - provider_type: remote::gemini - - provider_type: remote::vertexai - - provider_type: remote::groq - - provider_type: remote::sambanova - - provider_type: remote::azure - - provider_type: inline::sentence-transformers - vector_io: - - provider_type: inline::faiss - - provider_type: inline::sqlite-vec - - provider_type: inline::milvus - - provider_type: remote::chromadb - - provider_type: remote::pgvector - - provider_type: remote::qdrant - - provider_type: remote::weaviate - files: - - provider_type: inline::localfs - safety: - - provider_type: inline::llama-guard - - provider_type: inline::code-scanner - agents: - - provider_type: inline::meta-reference - post_training: - - provider_type: inline::torchtune-cpu - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol - batches: - - provider_type: inline::reference -image_type: venv diff --git a/src/llama_stack/distributions/dell/build.yaml b/src/llama_stack/distributions/dell/build.yaml deleted file mode 100644 index 2c809e0fe..000000000 --- a/src/llama_stack/distributions/dell/build.yaml +++ /dev/null @@ -1,30 +0,0 @@ -version: 2 -distribution_spec: - description: Dell's distribution of Llama Stack. TGI inference via Dell's custom - container - providers: - inference: - - provider_type: remote::tgi - - provider_type: inline::sentence-transformers - vector_io: - - provider_type: inline::faiss - - provider_type: remote::chromadb - - provider_type: remote::pgvector - safety: - - provider_type: inline::llama-guard - agents: - - provider_type: inline::meta-reference - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime -image_type: venv diff --git a/src/llama_stack/distributions/meta-reference-gpu/build.yaml b/src/llama_stack/distributions/meta-reference-gpu/build.yaml deleted file mode 100644 index 62a845b83..000000000 --- a/src/llama_stack/distributions/meta-reference-gpu/build.yaml +++ /dev/null @@ -1,29 +0,0 @@ -version: 2 -distribution_spec: - description: Use Meta Reference for running LLM inference - providers: - inference: - - provider_type: inline::meta-reference - vector_io: - - provider_type: inline::faiss - - provider_type: remote::chromadb - - provider_type: remote::pgvector - safety: - - provider_type: inline::llama-guard - agents: - - provider_type: inline::meta-reference - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol -image_type: venv diff --git a/src/llama_stack/distributions/nvidia/build.yaml b/src/llama_stack/distributions/nvidia/build.yaml deleted file mode 100644 index 86e085575..000000000 --- a/src/llama_stack/distributions/nvidia/build.yaml +++ /dev/null @@ -1,26 +0,0 @@ -version: 2 -distribution_spec: - description: Use NVIDIA NIM for running LLM inference, evaluation and safety - providers: - inference: - - provider_type: remote::nvidia - vector_io: - - provider_type: inline::faiss - safety: - - provider_type: remote::nvidia - agents: - - provider_type: inline::meta-reference - eval: - - provider_type: remote::nvidia - post_training: - - provider_type: remote::nvidia - datasetio: - - provider_type: inline::localfs - - provider_type: remote::nvidia - scoring: - - provider_type: inline::basic - tool_runtime: - - provider_type: inline::rag-runtime - files: - - provider_type: inline::localfs -image_type: venv diff --git a/src/llama_stack/distributions/oci/build.yaml b/src/llama_stack/distributions/oci/build.yaml deleted file mode 100644 index ac9fe967d..000000000 --- a/src/llama_stack/distributions/oci/build.yaml +++ /dev/null @@ -1,32 +0,0 @@ -version: 2 -distribution_spec: - description: Use Oracle Cloud Infrastructure (OCI) Generative AI for running LLM - inference with scalable cloud services - providers: - inference: - - provider_type: remote::oci - vector_io: - - provider_type: inline::faiss - - provider_type: remote::chromadb - - provider_type: remote::pgvector - safety: - - provider_type: inline::llama-guard - agents: - - provider_type: inline::meta-reference - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol - files: - - provider_type: inline::localfs -image_type: venv diff --git a/src/llama_stack/distributions/open-benchmark/build.yaml b/src/llama_stack/distributions/open-benchmark/build.yaml deleted file mode 100644 index 9613ef60d..000000000 --- a/src/llama_stack/distributions/open-benchmark/build.yaml +++ /dev/null @@ -1,33 +0,0 @@ -version: 2 -distribution_spec: - description: Distribution for running open benchmarks - providers: - inference: - - provider_type: remote::openai - - provider_type: remote::anthropic - - provider_type: remote::gemini - - provider_type: remote::groq - - provider_type: remote::together - vector_io: - - provider_type: inline::sqlite-vec - - provider_type: remote::chromadb - - provider_type: remote::pgvector - safety: - - provider_type: inline::llama-guard - agents: - - provider_type: inline::meta-reference - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol -image_type: venv diff --git a/src/llama_stack/distributions/starter-gpu/build.yaml b/src/llama_stack/distributions/starter-gpu/build.yaml deleted file mode 100644 index 57ba12dc4..000000000 --- a/src/llama_stack/distributions/starter-gpu/build.yaml +++ /dev/null @@ -1,56 +0,0 @@ -version: 2 -distribution_spec: - description: Quick start template for running Llama Stack with several popular providers. - This distribution is intended for GPU-enabled environments. - providers: - inference: - - provider_type: remote::cerebras - - provider_type: remote::ollama - - provider_type: remote::vllm - - provider_type: remote::tgi - - provider_type: remote::fireworks - - provider_type: remote::together - - provider_type: remote::bedrock - - provider_type: remote::nvidia - - provider_type: remote::openai - - provider_type: remote::anthropic - - provider_type: remote::gemini - - provider_type: remote::vertexai - - provider_type: remote::groq - - provider_type: remote::sambanova - - provider_type: remote::azure - - provider_type: inline::sentence-transformers - vector_io: - - provider_type: inline::faiss - - provider_type: inline::sqlite-vec - - provider_type: inline::milvus - - provider_type: remote::chromadb - - provider_type: remote::pgvector - - provider_type: remote::qdrant - - provider_type: remote::weaviate - files: - - provider_type: inline::localfs - safety: - - provider_type: inline::llama-guard - - provider_type: inline::code-scanner - agents: - - provider_type: inline::meta-reference - post_training: - - provider_type: inline::huggingface-gpu - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol - batches: - - provider_type: inline::reference -image_type: venv diff --git a/src/llama_stack/distributions/starter/build.yaml b/src/llama_stack/distributions/starter/build.yaml deleted file mode 100644 index 9fe7ccd14..000000000 --- a/src/llama_stack/distributions/starter/build.yaml +++ /dev/null @@ -1,56 +0,0 @@ -version: 2 -distribution_spec: - description: Quick start template for running Llama Stack with several popular providers. - This distribution is intended for CPU-only environments. - providers: - inference: - - provider_type: remote::cerebras - - provider_type: remote::ollama - - provider_type: remote::vllm - - provider_type: remote::tgi - - provider_type: remote::fireworks - - provider_type: remote::together - - provider_type: remote::bedrock - - provider_type: remote::nvidia - - provider_type: remote::openai - - provider_type: remote::anthropic - - provider_type: remote::gemini - - provider_type: remote::vertexai - - provider_type: remote::groq - - provider_type: remote::sambanova - - provider_type: remote::azure - - provider_type: inline::sentence-transformers - vector_io: - - provider_type: inline::faiss - - provider_type: inline::sqlite-vec - - provider_type: inline::milvus - - provider_type: remote::chromadb - - provider_type: remote::pgvector - - provider_type: remote::qdrant - - provider_type: remote::weaviate - files: - - provider_type: inline::localfs - safety: - - provider_type: inline::llama-guard - - provider_type: inline::code-scanner - agents: - - provider_type: inline::meta-reference - post_training: - - provider_type: inline::torchtune-cpu - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol - batches: - - provider_type: inline::reference -image_type: venv diff --git a/src/llama_stack/distributions/watsonx/build.yaml b/src/llama_stack/distributions/watsonx/build.yaml deleted file mode 100644 index 8fc27c8c8..000000000 --- a/src/llama_stack/distributions/watsonx/build.yaml +++ /dev/null @@ -1,30 +0,0 @@ -version: 2 -distribution_spec: - description: Use watsonx for running LLM inference - providers: - inference: - - provider_type: remote::watsonx - - provider_type: inline::sentence-transformers - vector_io: - - provider_type: inline::faiss - safety: - - provider_type: inline::llama-guard - agents: - - provider_type: inline::meta-reference - eval: - - provider_type: inline::meta-reference - datasetio: - - provider_type: remote::huggingface - - provider_type: inline::localfs - scoring: - - provider_type: inline::basic - - provider_type: inline::llm-as-judge - - provider_type: inline::braintrust - tool_runtime: - - provider_type: remote::brave-search - - provider_type: remote::tavily-search - - provider_type: inline::rag-runtime - - provider_type: remote::model-context-protocol - files: - - provider_type: inline::localfs -image_type: venv diff --git a/tests/external/build.yaml b/tests/external/build.yaml deleted file mode 100644 index fde6cb178..000000000 --- a/tests/external/build.yaml +++ /dev/null @@ -1,10 +0,0 @@ -version: '2' -distribution_spec: - description: Custom distro for CI tests - providers: - weather: - - provider_type: remote::kaze -image_type: venv -image_name: ci-test -external_providers_dir: ~/.llama/providers.d -external_apis_dir: ~/.llama/apis.d diff --git a/tests/external/ramalama-stack/build.yaml b/tests/external/ramalama-stack/build.yaml deleted file mode 100644 index 4403339f1..000000000 --- a/tests/external/ramalama-stack/build.yaml +++ /dev/null @@ -1,10 +0,0 @@ -version: 2 -distribution_spec: - description: Use (an external) Ramalama server for running LLM inference - container_image: null - providers: - inference: - - provider_type: remote::ramalama - module: ramalama_stack==0.3.0a0 -image_type: venv -image_name: ramalama-stack-test