diff --git a/llama_stack/templates/verification/__init__.py b/llama_stack/templates/verification/__init__.py deleted file mode 100644 index 5d8c281a6..000000000 --- a/llama_stack/templates/verification/__init__.py +++ /dev/null @@ -1,7 +0,0 @@ -# Copyright (c) Meta Platforms, Inc. and affiliates. -# All rights reserved. -# -# This source code is licensed under the terms described in the LICENSE file in -# the root directory of this source tree. - -from .verification import get_distribution_template # noqa: F401 diff --git a/llama_stack/templates/verification/build.yaml b/llama_stack/templates/verification/build.yaml deleted file mode 100644 index ce083dbba..000000000 --- a/llama_stack/templates/verification/build.yaml +++ /dev/null @@ -1,40 +0,0 @@ -version: '2' -distribution_spec: - description: Distribution for running e2e tests in CI - providers: - inference: - - remote::openai - - remote::fireworks-openai-compat - - remote::together-openai-compat - - remote::groq-openai-compat - - remote::sambanova-openai-compat - - remote::cerebras-openai-compat - - inline::sentence-transformers - vector_io: - - inline::sqlite-vec - - remote::chromadb - - remote::pgvector - safety: - - inline::llama-guard - agents: - - inline::meta-reference - telemetry: - - inline::meta-reference - eval: - - inline::meta-reference - datasetio: - - remote::huggingface - - inline::localfs - scoring: - - inline::basic - - inline::llm-as-judge - - inline::braintrust - tool_runtime: - - remote::brave-search - - remote::tavily-search - - inline::rag-runtime - - remote::model-context-protocol -image_type: conda -additional_pip_packages: -- aiosqlite -- sqlalchemy[asyncio] diff --git a/llama_stack/templates/verification/run.yaml b/llama_stack/templates/verification/run.yaml deleted file mode 100644 index 91e95bccc..000000000 --- a/llama_stack/templates/verification/run.yaml +++ /dev/null @@ -1,731 +0,0 @@ -version: '2' -image_name: verification -apis: -- agents -- datasetio -- eval -- inference -- safety -- scoring -- telemetry -- tool_runtime -- vector_io -providers: - inference: - - provider_id: openai - provider_type: remote::openai - config: - api_key: ${env.OPENAI_API_KEY:} - - provider_id: fireworks-openai-compat - provider_type: remote::fireworks-openai-compat - config: - openai_compat_api_base: https://api.fireworks.ai/inference/v1 - api_key: ${env.FIREWORKS_API_KEY:} - - provider_id: together-openai-compat - provider_type: remote::together-openai-compat - config: - openai_compat_api_base: https://api.together.xyz/v1 - api_key: ${env.TOGETHER_API_KEY:} - - provider_id: groq-openai-compat - provider_type: remote::groq-openai-compat - config: - openai_compat_api_base: https://api.groq.com/openai/v1 - api_key: ${env.GROQ_API_KEY:} - - provider_id: sambanova-openai-compat - provider_type: remote::sambanova-openai-compat - config: - openai_compat_api_base: https://api.sambanova.ai/v1 - api_key: ${env.SAMBANOVA_API_KEY:} - - provider_id: cerebras-openai-compat - provider_type: remote::cerebras-openai-compat - config: - openai_compat_api_base: https://api.cerebras.ai/v1 - api_key: ${env.CEREBRAS_API_KEY:} - - provider_id: sentence-transformers - provider_type: inline::sentence-transformers - config: {} - vector_io: - - provider_id: sqlite-vec - provider_type: inline::sqlite-vec - config: - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/sqlite_vec.db - - provider_id: ${env.ENABLE_CHROMADB+chromadb} - provider_type: remote::chromadb - config: - url: ${env.CHROMADB_URL:} - - provider_id: ${env.ENABLE_PGVECTOR+pgvector} - provider_type: remote::pgvector - config: - host: ${env.PGVECTOR_HOST:localhost} - port: ${env.PGVECTOR_PORT:5432} - db: ${env.PGVECTOR_DB:} - user: ${env.PGVECTOR_USER:} - password: ${env.PGVECTOR_PASSWORD:} - safety: - - provider_id: llama-guard - provider_type: inline::llama-guard - config: - excluded_categories: [] - agents: - - provider_id: meta-reference - provider_type: inline::meta-reference - config: - persistence_store: - type: sqlite - namespace: null - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/agents_store.db - responses_store: - type: sqlite - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/responses_store.db - telemetry: - - provider_id: meta-reference - provider_type: inline::meta-reference - config: - service_name: "${env.OTEL_SERVICE_NAME:\u200B}" - sinks: ${env.TELEMETRY_SINKS:console,sqlite} - sqlite_db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/trace_store.db - eval: - - provider_id: meta-reference - provider_type: inline::meta-reference - config: - kvstore: - type: sqlite - namespace: null - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/meta_reference_eval.db - datasetio: - - provider_id: huggingface - provider_type: remote::huggingface - config: - kvstore: - type: sqlite - namespace: null - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/huggingface_datasetio.db - - provider_id: localfs - provider_type: inline::localfs - config: - kvstore: - type: sqlite - namespace: null - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/localfs_datasetio.db - scoring: - - provider_id: basic - provider_type: inline::basic - config: {} - - provider_id: llm-as-judge - provider_type: inline::llm-as-judge - config: {} - - provider_id: braintrust - provider_type: inline::braintrust - config: - openai_api_key: ${env.OPENAI_API_KEY:} - tool_runtime: - - provider_id: brave-search - provider_type: remote::brave-search - config: - api_key: ${env.BRAVE_SEARCH_API_KEY:} - max_results: 3 - - provider_id: tavily-search - provider_type: remote::tavily-search - config: - api_key: ${env.TAVILY_SEARCH_API_KEY:} - max_results: 3 - - provider_id: rag-runtime - provider_type: inline::rag-runtime - config: {} - - provider_id: model-context-protocol - provider_type: remote::model-context-protocol - config: {} -metadata_store: - type: sqlite - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/registry.db -inference_store: - type: sqlite - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/verification}/inference_store.db -models: -- metadata: {} - model_id: openai/gpt-4o - provider_id: openai - provider_model_id: openai/gpt-4o - model_type: llm -- metadata: {} - model_id: openai/gpt-4o-mini - provider_id: openai - provider_model_id: openai/gpt-4o-mini - model_type: llm -- metadata: {} - model_id: openai/chatgpt-4o-latest - provider_id: openai - provider_model_id: openai/chatgpt-4o-latest - model_type: llm -- metadata: {} - model_id: openai/gpt-3.5-turbo-0125 - provider_id: openai - provider_model_id: gpt-3.5-turbo-0125 - model_type: llm -- metadata: {} - model_id: openai/gpt-3.5-turbo - provider_id: openai - provider_model_id: gpt-3.5-turbo - model_type: llm -- metadata: {} - model_id: openai/gpt-3.5-turbo-instruct - provider_id: openai - provider_model_id: gpt-3.5-turbo-instruct - model_type: llm -- metadata: {} - model_id: openai/gpt-4 - provider_id: openai - provider_model_id: gpt-4 - model_type: llm -- metadata: {} - model_id: openai/gpt-4-turbo - provider_id: openai - provider_model_id: gpt-4-turbo - model_type: llm -- metadata: {} - model_id: openai/gpt-4o - provider_id: openai - provider_model_id: gpt-4o - model_type: llm -- metadata: {} - model_id: openai/gpt-4o-2024-08-06 - provider_id: openai - provider_model_id: gpt-4o-2024-08-06 - model_type: llm -- metadata: {} - model_id: openai/gpt-4o-mini - provider_id: openai - provider_model_id: gpt-4o-mini - model_type: llm -- metadata: {} - model_id: openai/gpt-4o-audio-preview - provider_id: openai - provider_model_id: gpt-4o-audio-preview - model_type: llm -- metadata: {} - model_id: openai/chatgpt-4o-latest - provider_id: openai - provider_model_id: chatgpt-4o-latest - model_type: llm -- metadata: {} - model_id: openai/o1 - provider_id: openai - provider_model_id: o1 - model_type: llm -- metadata: {} - model_id: openai/o1-mini - provider_id: openai - provider_model_id: o1-mini - model_type: llm -- metadata: {} - model_id: openai/o3-mini - provider_id: openai - provider_model_id: o3-mini - model_type: llm -- metadata: {} - model_id: openai/o4-mini - provider_id: openai - provider_model_id: o4-mini - model_type: llm -- metadata: - embedding_dimension: 1536 - context_length: 8192 - model_id: openai/text-embedding-3-small - provider_id: openai - provider_model_id: openai/text-embedding-3-small - model_type: embedding -- metadata: - embedding_dimension: 3072 - context_length: 8192 - model_id: openai/text-embedding-3-large - provider_id: openai - provider_model_id: openai/text-embedding-3-large - model_type: embedding -- metadata: - embedding_dimension: 1536 - context_length: 8192 - model_id: openai/text-embedding-3-small - provider_id: openai - provider_model_id: text-embedding-3-small - model_type: embedding -- metadata: - embedding_dimension: 3072 - context_length: 8192 - model_id: openai/text-embedding-3-large - provider_id: openai - provider_model_id: text-embedding-3-large - model_type: embedding -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p1-8b-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p1-8b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.1-8B-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p1-8b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p1-70b-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p1-70b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.1-70B-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p1-70b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p1-405b-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p1-405b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.1-405B-Instruct-FP8 - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p1-405b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p2-3b-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p2-3b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.2-3B-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p2-3b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p2-11b-vision-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p2-11b-vision-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.2-11B-Vision-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p2-11b-vision-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p2-90b-vision-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p2-90b-vision-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.2-90B-Vision-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p2-90b-vision-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-v3p3-70b-instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p3-70b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-3.3-70B-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-v3p3-70b-instruct - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-guard-3-8b - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-guard-3-8b - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-Guard-3-8B - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-guard-3-8b - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama-guard-3-11b-vision - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-guard-3-11b-vision - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-Guard-3-11B-Vision - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama-guard-3-11b-vision - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama4-scout-instruct-basic - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama4-scout-instruct-basic - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama4-scout-instruct-basic - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/accounts/fireworks/models/llama4-maverick-instruct-basic - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama4-maverick-instruct-basic - model_type: llm -- metadata: {} - model_id: fireworks-openai-compat/meta-llama/Llama-4-Maverick-17B-128E-Instruct - provider_id: fireworks-openai-compat - provider_model_id: accounts/fireworks/models/llama4-maverick-instruct-basic - model_type: llm -- metadata: - embedding_dimension: 768 - context_length: 8192 - model_id: fireworks-openai-compat/nomic-ai/nomic-embed-text-v1.5 - provider_id: fireworks-openai-compat - provider_model_id: nomic-ai/nomic-embed-text-v1.5 - model_type: embedding -- metadata: {} - model_id: together-openai-compat/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.1-8B-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.1-70B-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.1-405B-Instruct-FP8 - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.2-3B-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.2-3B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.2-3B-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.2-3B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.2-11B-Vision-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.2-90B-Vision-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.3-70B-Instruct-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.3-70B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-3.3-70B-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-3.3-70B-Instruct-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Meta-Llama-Guard-3-8B - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-Guard-3-8B - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-Guard-3-8B - provider_id: together-openai-compat - provider_model_id: meta-llama/Meta-Llama-Guard-3-8B - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-Guard-3-11B-Vision-Turbo - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-Guard-3-11B-Vision-Turbo - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-Guard-3-11B-Vision - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-Guard-3-11B-Vision-Turbo - model_type: llm -- metadata: - embedding_dimension: 768 - context_length: 8192 - model_id: together-openai-compat/togethercomputer/m2-bert-80M-8k-retrieval - provider_id: together-openai-compat - provider_model_id: togethercomputer/m2-bert-80M-8k-retrieval - model_type: embedding -- metadata: - embedding_dimension: 768 - context_length: 32768 - model_id: together-openai-compat/togethercomputer/m2-bert-80M-32k-retrieval - provider_id: together-openai-compat - provider_model_id: togethercomputer/m2-bert-80M-32k-retrieval - model_type: embedding -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-4-Scout-17B-16E-Instruct - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-4-Scout-17B-16E-Instruct - model_type: llm -- metadata: {} - model_id: together-openai-compat/together/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-4-Scout-17B-16E-Instruct - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 - model_type: llm -- metadata: {} - model_id: together-openai-compat/meta-llama/Llama-4-Maverick-17B-128E-Instruct - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 - model_type: llm -- metadata: {} - model_id: together-openai-compat/together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 - provider_id: together-openai-compat - provider_model_id: meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama3-8b-8192 - provider_id: groq-openai-compat - provider_model_id: groq/llama3-8b-8192 - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-3.1-8B-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama3-8b-8192 - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama-3.1-8b-instant - provider_id: groq-openai-compat - provider_model_id: groq/llama-3.1-8b-instant - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama3-70b-8192 - provider_id: groq-openai-compat - provider_model_id: groq/llama3-70b-8192 - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-3-70B-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama3-70b-8192 - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama-3.3-70b-versatile - provider_id: groq-openai-compat - provider_model_id: groq/llama-3.3-70b-versatile - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-3.3-70B-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama-3.3-70b-versatile - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama-3.2-3b-preview - provider_id: groq-openai-compat - provider_model_id: groq/llama-3.2-3b-preview - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-3.2-3B-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama-3.2-3b-preview - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama-4-scout-17b-16e-instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama-4-scout-17b-16e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama-4-scout-17b-16e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/meta-llama/llama-4-scout-17b-16e-instruct - provider_id: groq-openai-compat - provider_model_id: groq/meta-llama/llama-4-scout-17b-16e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/meta-llama/llama-4-scout-17b-16e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/llama-4-maverick-17b-128e-instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama-4-maverick-17b-128e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-4-Maverick-17B-128E-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/llama-4-maverick-17b-128e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/groq/meta-llama/llama-4-maverick-17b-128e-instruct - provider_id: groq-openai-compat - provider_model_id: groq/meta-llama/llama-4-maverick-17b-128e-instruct - model_type: llm -- metadata: {} - model_id: groq-openai-compat/meta-llama/Llama-4-Maverick-17B-128E-Instruct - provider_id: groq-openai-compat - provider_model_id: groq/meta-llama/llama-4-maverick-17b-128e-instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Meta-Llama-3.1-8B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.1-8B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.1-8B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.1-8B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Meta-Llama-3.1-405B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.1-405B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.1-405B-Instruct-FP8 - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.1-405B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Meta-Llama-3.2-1B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.2-1B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.2-1B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.2-1B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Meta-Llama-3.2-3B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.2-3B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.2-3B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.2-3B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Meta-Llama-3.3-70B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.3-70B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.3-70B-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-3.3-70B-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Llama-3.2-11B-Vision-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-3.2-11B-Vision-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.2-11B-Vision-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-3.2-11B-Vision-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Llama-3.2-90B-Vision-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-3.2-90B-Vision-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-3.2-90B-Vision-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-3.2-90B-Vision-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Llama-4-Scout-17B-16E-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-4-Scout-17B-16E-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-4-Scout-17B-16E-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-4-Scout-17B-16E-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Llama-4-Maverick-17B-128E-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-4-Maverick-17B-128E-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-4-Maverick-17B-128E-Instruct - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Llama-4-Maverick-17B-128E-Instruct - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/sambanova/Meta-Llama-Guard-3-8B - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-Guard-3-8B - model_type: llm -- metadata: {} - model_id: sambanova-openai-compat/meta-llama/Llama-Guard-3-8B - provider_id: sambanova-openai-compat - provider_model_id: sambanova/Meta-Llama-Guard-3-8B - model_type: llm -- metadata: {} - model_id: cerebras-openai-compat/llama3.1-8b - provider_id: cerebras-openai-compat - provider_model_id: llama3.1-8b - model_type: llm -- metadata: {} - model_id: cerebras-openai-compat/meta-llama/Llama-3.1-8B-Instruct - provider_id: cerebras-openai-compat - provider_model_id: llama3.1-8b - model_type: llm -- metadata: {} - model_id: cerebras-openai-compat/llama-3.3-70b - provider_id: cerebras-openai-compat - provider_model_id: llama-3.3-70b - model_type: llm -- metadata: {} - model_id: cerebras-openai-compat/meta-llama/Llama-3.3-70B-Instruct - provider_id: cerebras-openai-compat - provider_model_id: llama-3.3-70b - model_type: llm -- metadata: - embedding_dimension: 384 - model_id: all-MiniLM-L6-v2 - provider_id: sentence-transformers - model_type: embedding -shields: -- shield_id: meta-llama/Llama-Guard-3-8B -vector_dbs: [] -datasets: [] -scoring_fns: [] -benchmarks: [] -tool_groups: -- toolgroup_id: builtin::websearch - provider_id: tavily-search -- toolgroup_id: builtin::rag - provider_id: rag-runtime -server: - port: 8321 diff --git a/llama_stack/templates/verification/verification.py b/llama_stack/templates/verification/verification.py deleted file mode 100644 index b58400f26..000000000 --- a/llama_stack/templates/verification/verification.py +++ /dev/null @@ -1,201 +0,0 @@ -# Copyright (c) Meta Platforms, Inc. and affiliates. -# All rights reserved. -# -# This source code is licensed under the terms described in the LICENSE file in -# the root directory of this source tree. - - -from llama_stack.apis.models.models import ModelType -from llama_stack.distribution.datatypes import ( - ModelInput, - Provider, - ShieldInput, - ToolGroupInput, -) -from llama_stack.providers.inline.inference.sentence_transformers import ( - SentenceTransformersInferenceConfig, -) -from llama_stack.providers.inline.vector_io.sqlite_vec.config import ( - SQLiteVectorIOConfig, -) -from llama_stack.providers.remote.inference.cerebras.models import MODEL_ENTRIES as CEREBRAS_MODEL_ENTRIES -from llama_stack.providers.remote.inference.cerebras_openai_compat.config import CerebrasCompatConfig -from llama_stack.providers.remote.inference.fireworks.models import ( - MODEL_ENTRIES as FIREWORKS_MODEL_ENTRIES, -) -from llama_stack.providers.remote.inference.fireworks_openai_compat.config import FireworksCompatConfig -from llama_stack.providers.remote.inference.groq.models import ( - MODEL_ENTRIES as GROQ_MODEL_ENTRIES, -) -from llama_stack.providers.remote.inference.groq_openai_compat.config import GroqCompatConfig -from llama_stack.providers.remote.inference.openai.config import OpenAIConfig -from llama_stack.providers.remote.inference.openai.models import ( - MODEL_ENTRIES as OPENAI_MODEL_ENTRIES, -) -from llama_stack.providers.remote.inference.sambanova.models import MODEL_ENTRIES as SAMBANOVA_MODEL_ENTRIES -from llama_stack.providers.remote.inference.sambanova_openai_compat.config import SambaNovaCompatConfig -from llama_stack.providers.remote.inference.together.models import ( - MODEL_ENTRIES as TOGETHER_MODEL_ENTRIES, -) -from llama_stack.providers.remote.inference.together_openai_compat.config import TogetherCompatConfig -from llama_stack.providers.remote.vector_io.chroma.config import ChromaVectorIOConfig -from llama_stack.providers.remote.vector_io.pgvector.config import ( - PGVectorVectorIOConfig, -) -from llama_stack.providers.utils.inference.model_registry import ProviderModelEntry -from llama_stack.templates.template import ( - DistributionTemplate, - RunConfigSettings, - get_model_registry, -) - - -def get_inference_providers() -> tuple[list[Provider], dict[str, list[ProviderModelEntry]]]: - # in this template, we allow each API key to be optional - providers = [ - ( - "openai", - OPENAI_MODEL_ENTRIES, - OpenAIConfig.sample_run_config(api_key="${env.OPENAI_API_KEY:}"), - ), - ( - "fireworks-openai-compat", - FIREWORKS_MODEL_ENTRIES, - FireworksCompatConfig.sample_run_config(api_key="${env.FIREWORKS_API_KEY:}"), - ), - ( - "together-openai-compat", - TOGETHER_MODEL_ENTRIES, - TogetherCompatConfig.sample_run_config(api_key="${env.TOGETHER_API_KEY:}"), - ), - ( - "groq-openai-compat", - GROQ_MODEL_ENTRIES, - GroqCompatConfig.sample_run_config(api_key="${env.GROQ_API_KEY:}"), - ), - ( - "sambanova-openai-compat", - SAMBANOVA_MODEL_ENTRIES, - SambaNovaCompatConfig.sample_run_config(api_key="${env.SAMBANOVA_API_KEY:}"), - ), - ( - "cerebras-openai-compat", - CEREBRAS_MODEL_ENTRIES, - CerebrasCompatConfig.sample_run_config(api_key="${env.CEREBRAS_API_KEY:}"), - ), - ] - inference_providers = [] - available_models = {} - for provider_id, model_entries, config in providers: - inference_providers.append( - Provider( - provider_id=provider_id, - provider_type=f"remote::{provider_id}", - config=config, - ) - ) - available_models[provider_id] = model_entries - return inference_providers, available_models - - -def get_distribution_template() -> DistributionTemplate: - inference_providers, available_models = get_inference_providers() - providers = { - "inference": ([p.provider_type for p in inference_providers] + ["inline::sentence-transformers"]), - "vector_io": ["inline::sqlite-vec", "remote::chromadb", "remote::pgvector"], - "safety": ["inline::llama-guard"], - "agents": ["inline::meta-reference"], - "telemetry": ["inline::meta-reference"], - "eval": ["inline::meta-reference"], - "datasetio": ["remote::huggingface", "inline::localfs"], - "scoring": ["inline::basic", "inline::llm-as-judge", "inline::braintrust"], - "tool_runtime": [ - "remote::brave-search", - "remote::tavily-search", - "inline::rag-runtime", - "remote::model-context-protocol", - ], - } - name = "verification" - - vector_io_providers = [ - Provider( - provider_id="sqlite-vec", - provider_type="inline::sqlite-vec", - config=SQLiteVectorIOConfig.sample_run_config(f"~/.llama/distributions/{name}"), - ), - Provider( - provider_id="${env.ENABLE_CHROMADB+chromadb}", - provider_type="remote::chromadb", - config=ChromaVectorIOConfig.sample_run_config(url="${env.CHROMADB_URL:}"), - ), - Provider( - provider_id="${env.ENABLE_PGVECTOR+pgvector}", - provider_type="remote::pgvector", - config=PGVectorVectorIOConfig.sample_run_config( - db="${env.PGVECTOR_DB:}", - user="${env.PGVECTOR_USER:}", - password="${env.PGVECTOR_PASSWORD:}", - ), - ), - ] - embedding_provider = Provider( - provider_id="sentence-transformers", - provider_type="inline::sentence-transformers", - config=SentenceTransformersInferenceConfig.sample_run_config(), - ) - - default_tool_groups = [ - ToolGroupInput( - toolgroup_id="builtin::websearch", - provider_id="tavily-search", - ), - ToolGroupInput( - toolgroup_id="builtin::rag", - provider_id="rag-runtime", - ), - ] - embedding_model = ModelInput( - model_id="all-MiniLM-L6-v2", - provider_id=embedding_provider.provider_id, - model_type=ModelType.embedding, - metadata={ - "embedding_dimension": 384, - }, - ) - - default_models = get_model_registry(available_models) - return DistributionTemplate( - name=name, - distro_type="self_hosted", - description="Distribution for running e2e tests in CI", - container_image=None, - template_path=None, - providers=providers, - available_models_by_provider=available_models, - run_configs={ - "run.yaml": RunConfigSettings( - provider_overrides={ - "inference": inference_providers + [embedding_provider], - "vector_io": vector_io_providers, - }, - default_models=default_models + [embedding_model], - default_tool_groups=default_tool_groups, - default_shields=[ShieldInput(shield_id="meta-llama/Llama-Guard-3-8B")], - ), - }, - run_config_env_vars={ - "LLAMA_STACK_PORT": ( - "8321", - "Port for the Llama Stack distribution server", - ), - "FIREWORKS_API_KEY": ( - "", - "Fireworks API Key", - ), - "OPENAI_API_KEY": ( - "", - "OpenAI API Key", - ), - }, - )