mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
more fixes to postgres-store run yaml ugh
This commit is contained in:
parent
e6de865bcb
commit
7ce0c5c5dc
4 changed files with 89 additions and 101 deletions
|
|
@ -170,7 +170,7 @@ providers:
|
||||||
namespace: agents
|
namespace: agents
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
responses:
|
responses:
|
||||||
table_name: agent_responses
|
table_name: responses
|
||||||
backend: sql_default
|
backend: sql_default
|
||||||
max_write_queue_size: 10000
|
max_write_queue_size: 10000
|
||||||
num_writers: 4
|
num_writers: 4
|
||||||
|
|
@ -230,8 +230,6 @@ providers:
|
||||||
kvstore:
|
kvstore:
|
||||||
namespace: batches
|
namespace: batches
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
max_concurrent_batches: 1
|
|
||||||
max_concurrent_requests_per_batch: 10
|
|
||||||
storage:
|
storage:
|
||||||
backends:
|
backends:
|
||||||
kv_default:
|
kv_default:
|
||||||
|
|
@ -266,13 +264,30 @@ storage:
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
registered_resources:
|
registered_resources:
|
||||||
models: []
|
models: []
|
||||||
shields: []
|
shields:
|
||||||
|
- shield_id: llama-guard
|
||||||
|
provider_id: ${env.SAFETY_MODEL:+llama-guard}
|
||||||
|
provider_shield_id: ${env.SAFETY_MODEL:=}
|
||||||
|
- shield_id: code-scanner
|
||||||
|
provider_id: ${env.CODE_SCANNER_MODEL:+code-scanner}
|
||||||
|
provider_shield_id: ${env.CODE_SCANNER_MODEL:=}
|
||||||
vector_dbs: []
|
vector_dbs: []
|
||||||
datasets: []
|
datasets: []
|
||||||
scoring_fns: []
|
scoring_fns: []
|
||||||
benchmarks: []
|
benchmarks: []
|
||||||
tool_groups: []
|
tool_groups:
|
||||||
|
- toolgroup_id: builtin::websearch
|
||||||
|
provider_id: tavily-search
|
||||||
|
- toolgroup_id: builtin::rag
|
||||||
|
provider_id: rag-runtime
|
||||||
server:
|
server:
|
||||||
port: 8321
|
port: 8321
|
||||||
telemetry:
|
telemetry:
|
||||||
enabled: true
|
enabled: true
|
||||||
|
vector_stores:
|
||||||
|
default_provider_id: faiss
|
||||||
|
default_embedding_model:
|
||||||
|
provider_id: sentence-transformers
|
||||||
|
model_id: nomic-ai/nomic-embed-text-v1.5
|
||||||
|
safety:
|
||||||
|
default_shield_id: llama-guard
|
||||||
|
|
|
||||||
|
|
@ -170,7 +170,7 @@ providers:
|
||||||
namespace: agents
|
namespace: agents
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
responses:
|
responses:
|
||||||
table_name: agent_responses
|
table_name: responses
|
||||||
backend: sql_default
|
backend: sql_default
|
||||||
max_write_queue_size: 10000
|
max_write_queue_size: 10000
|
||||||
num_writers: 4
|
num_writers: 4
|
||||||
|
|
@ -233,8 +233,6 @@ providers:
|
||||||
kvstore:
|
kvstore:
|
||||||
namespace: batches
|
namespace: batches
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
max_concurrent_batches: 1
|
|
||||||
max_concurrent_requests_per_batch: 10
|
|
||||||
storage:
|
storage:
|
||||||
backends:
|
backends:
|
||||||
kv_default:
|
kv_default:
|
||||||
|
|
@ -269,13 +267,30 @@ storage:
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
registered_resources:
|
registered_resources:
|
||||||
models: []
|
models: []
|
||||||
shields: []
|
shields:
|
||||||
|
- shield_id: llama-guard
|
||||||
|
provider_id: ${env.SAFETY_MODEL:+llama-guard}
|
||||||
|
provider_shield_id: ${env.SAFETY_MODEL:=}
|
||||||
|
- shield_id: code-scanner
|
||||||
|
provider_id: ${env.CODE_SCANNER_MODEL:+code-scanner}
|
||||||
|
provider_shield_id: ${env.CODE_SCANNER_MODEL:=}
|
||||||
vector_dbs: []
|
vector_dbs: []
|
||||||
datasets: []
|
datasets: []
|
||||||
scoring_fns: []
|
scoring_fns: []
|
||||||
benchmarks: []
|
benchmarks: []
|
||||||
tool_groups: []
|
tool_groups:
|
||||||
|
- toolgroup_id: builtin::websearch
|
||||||
|
provider_id: tavily-search
|
||||||
|
- toolgroup_id: builtin::rag
|
||||||
|
provider_id: rag-runtime
|
||||||
server:
|
server:
|
||||||
port: 8321
|
port: 8321
|
||||||
telemetry:
|
telemetry:
|
||||||
enabled: true
|
enabled: true
|
||||||
|
vector_stores:
|
||||||
|
default_provider_id: faiss
|
||||||
|
default_embedding_model:
|
||||||
|
provider_id: sentence-transformers
|
||||||
|
model_id: nomic-ai/nomic-embed-text-v1.5
|
||||||
|
safety:
|
||||||
|
default_shield_id: llama-guard
|
||||||
|
|
|
||||||
|
|
@ -170,7 +170,7 @@ providers:
|
||||||
namespace: agents
|
namespace: agents
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
responses:
|
responses:
|
||||||
table_name: agent_responses
|
table_name: responses
|
||||||
backend: sql_default
|
backend: sql_default
|
||||||
max_write_queue_size: 10000
|
max_write_queue_size: 10000
|
||||||
num_writers: 4
|
num_writers: 4
|
||||||
|
|
@ -230,8 +230,6 @@ providers:
|
||||||
kvstore:
|
kvstore:
|
||||||
namespace: batches
|
namespace: batches
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
max_concurrent_batches: 1
|
|
||||||
max_concurrent_requests_per_batch: 10
|
|
||||||
storage:
|
storage:
|
||||||
backends:
|
backends:
|
||||||
kv_default:
|
kv_default:
|
||||||
|
|
@ -266,13 +264,30 @@ storage:
|
||||||
backend: kv_default
|
backend: kv_default
|
||||||
registered_resources:
|
registered_resources:
|
||||||
models: []
|
models: []
|
||||||
shields: []
|
shields:
|
||||||
|
- shield_id: llama-guard
|
||||||
|
provider_id: ${env.SAFETY_MODEL:+llama-guard}
|
||||||
|
provider_shield_id: ${env.SAFETY_MODEL:=}
|
||||||
|
- shield_id: code-scanner
|
||||||
|
provider_id: ${env.CODE_SCANNER_MODEL:+code-scanner}
|
||||||
|
provider_shield_id: ${env.CODE_SCANNER_MODEL:=}
|
||||||
vector_dbs: []
|
vector_dbs: []
|
||||||
datasets: []
|
datasets: []
|
||||||
scoring_fns: []
|
scoring_fns: []
|
||||||
benchmarks: []
|
benchmarks: []
|
||||||
tool_groups: []
|
tool_groups:
|
||||||
|
- toolgroup_id: builtin::websearch
|
||||||
|
provider_id: tavily-search
|
||||||
|
- toolgroup_id: builtin::rag
|
||||||
|
provider_id: rag-runtime
|
||||||
server:
|
server:
|
||||||
port: 8321
|
port: 8321
|
||||||
telemetry:
|
telemetry:
|
||||||
enabled: true
|
enabled: true
|
||||||
|
vector_stores:
|
||||||
|
default_provider_id: faiss
|
||||||
|
default_embedding_model:
|
||||||
|
provider_id: sentence-transformers
|
||||||
|
model_id: nomic-ai/nomic-embed-text-v1.5
|
||||||
|
safety:
|
||||||
|
default_shield_id: llama-guard
|
||||||
|
|
|
||||||
|
|
@ -17,22 +17,9 @@ from llama_stack.core.datatypes import (
|
||||||
ToolGroupInput,
|
ToolGroupInput,
|
||||||
VectorStoresConfig,
|
VectorStoresConfig,
|
||||||
)
|
)
|
||||||
from llama_stack.core.storage.datatypes import (
|
|
||||||
InferenceStoreReference,
|
|
||||||
KVStoreReference,
|
|
||||||
ResponsesStoreReference,
|
|
||||||
SqlStoreReference,
|
|
||||||
)
|
|
||||||
from llama_stack.core.utils.dynamic import instantiate_class_type
|
from llama_stack.core.utils.dynamic import instantiate_class_type
|
||||||
from llama_stack.distributions.template import DistributionTemplate, RunConfigSettings
|
from llama_stack.distributions.template import DistributionTemplate, RunConfigSettings
|
||||||
from llama_stack.providers.datatypes import RemoteProviderSpec
|
from llama_stack.providers.datatypes import RemoteProviderSpec
|
||||||
from llama_stack.providers.inline.agents.meta_reference.config import (
|
|
||||||
AgentPersistenceConfig,
|
|
||||||
MetaReferenceAgentsImplConfig,
|
|
||||||
)
|
|
||||||
from llama_stack.providers.inline.batches.reference.config import (
|
|
||||||
ReferenceBatchesImplConfig,
|
|
||||||
)
|
|
||||||
from llama_stack.providers.inline.files.localfs.config import LocalfsFilesImplConfig
|
from llama_stack.providers.inline.files.localfs.config import LocalfsFilesImplConfig
|
||||||
from llama_stack.providers.inline.inference.sentence_transformers import (
|
from llama_stack.providers.inline.inference.sentence_transformers import (
|
||||||
SentenceTransformersInferenceConfig,
|
SentenceTransformersInferenceConfig,
|
||||||
|
|
@ -254,16 +241,7 @@ def get_distribution_template(name: str = "starter") -> DistributionTemplate:
|
||||||
"files": [files_provider],
|
"files": [files_provider],
|
||||||
}
|
}
|
||||||
|
|
||||||
return DistributionTemplate(
|
base_run_settings = RunConfigSettings(
|
||||||
name=name,
|
|
||||||
distro_type="self_hosted",
|
|
||||||
description="Quick start template for running Llama Stack with several popular providers. This distribution is intended for CPU-only environments.",
|
|
||||||
container_image=None,
|
|
||||||
template_path=None,
|
|
||||||
providers=providers,
|
|
||||||
additional_pip_packages=list(set(PostgresSqlStoreConfig.pip_packages() + PostgresKVStoreConfig.pip_packages())),
|
|
||||||
run_configs={
|
|
||||||
"run.yaml": RunConfigSettings(
|
|
||||||
provider_overrides=default_overrides,
|
provider_overrides=default_overrides,
|
||||||
default_models=[],
|
default_models=[],
|
||||||
default_tool_groups=default_tool_groups,
|
default_tool_groups=default_tool_groups,
|
||||||
|
|
@ -278,64 +256,29 @@ def get_distribution_template(name: str = "starter") -> DistributionTemplate:
|
||||||
safety_config=SafetyConfig(
|
safety_config=SafetyConfig(
|
||||||
default_shield_id="llama-guard",
|
default_shield_id="llama-guard",
|
||||||
),
|
),
|
||||||
),
|
|
||||||
"run-with-postgres-store.yaml": RunConfigSettings(
|
|
||||||
provider_overrides={
|
|
||||||
**default_overrides,
|
|
||||||
"agents": [
|
|
||||||
Provider(
|
|
||||||
provider_id="meta-reference",
|
|
||||||
provider_type="inline::meta-reference",
|
|
||||||
config=MetaReferenceAgentsImplConfig(
|
|
||||||
persistence=AgentPersistenceConfig(
|
|
||||||
agent_state=KVStoreReference(
|
|
||||||
backend="kv_default",
|
|
||||||
namespace="agents",
|
|
||||||
),
|
|
||||||
responses=ResponsesStoreReference(
|
|
||||||
backend="sql_default",
|
|
||||||
table_name="agent_responses",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
).model_dump(exclude_none=True),
|
|
||||||
)
|
)
|
||||||
],
|
|
||||||
"batches": [
|
postgres_run_settings = base_run_settings.model_copy(
|
||||||
Provider(
|
update={
|
||||||
provider_id="reference",
|
"storage_backends": {
|
||||||
provider_type="inline::reference",
|
|
||||||
config=ReferenceBatchesImplConfig(
|
|
||||||
kvstore=KVStoreReference(
|
|
||||||
backend="kv_default",
|
|
||||||
namespace="batches",
|
|
||||||
),
|
|
||||||
).model_dump(exclude_none=True),
|
|
||||||
)
|
|
||||||
],
|
|
||||||
},
|
|
||||||
storage_backends={
|
|
||||||
"kv_default": postgres_kv_config,
|
"kv_default": postgres_kv_config,
|
||||||
"sql_default": postgres_sql_config,
|
"sql_default": postgres_sql_config,
|
||||||
|
}
|
||||||
},
|
},
|
||||||
storage_stores={
|
deep=True,
|
||||||
"metadata": KVStoreReference(
|
)
|
||||||
backend="kv_default",
|
|
||||||
namespace="registry",
|
return DistributionTemplate(
|
||||||
).model_dump(exclude_none=True),
|
name=name,
|
||||||
"inference": InferenceStoreReference(
|
distro_type="self_hosted",
|
||||||
backend="sql_default",
|
description="Quick start template for running Llama Stack with several popular providers. This distribution is intended for CPU-only environments.",
|
||||||
table_name="inference_store",
|
container_image=None,
|
||||||
).model_dump(exclude_none=True),
|
template_path=None,
|
||||||
"conversations": SqlStoreReference(
|
providers=providers,
|
||||||
backend="sql_default",
|
additional_pip_packages=list(set(PostgresSqlStoreConfig.pip_packages() + PostgresKVStoreConfig.pip_packages())),
|
||||||
table_name="openai_conversations",
|
run_configs={
|
||||||
).model_dump(exclude_none=True),
|
"run.yaml": base_run_settings,
|
||||||
"prompts": KVStoreReference(
|
"run-with-postgres-store.yaml": postgres_run_settings,
|
||||||
backend="kv_default",
|
|
||||||
namespace="prompts",
|
|
||||||
).model_dump(exclude_none=True),
|
|
||||||
},
|
|
||||||
),
|
|
||||||
},
|
},
|
||||||
run_config_env_vars={
|
run_config_env_vars={
|
||||||
"LLAMA_STACK_PORT": (
|
"LLAMA_STACK_PORT": (
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue