undo SafetyConfig, fix pre-commit

This commit is contained in:
Ashwin Bharambe 2025-11-12 12:48:03 -08:00
parent 81e44b06ff
commit 283a998137
18 changed files with 585 additions and 681 deletions

View file

@ -79,33 +79,6 @@ docker run \
--port $LLAMA_STACK_PORT --port $LLAMA_STACK_PORT
``` ```
### Via Docker with Custom Run Configuration
You can also run the Docker container with a custom run configuration file by mounting it into the container:
```bash
# Set the path to your custom run.yaml file
CUSTOM_RUN_CONFIG=/path/to/your/custom-run.yaml
LLAMA_STACK_PORT=8321
docker run \
-it \
--pull always \
--gpu all \
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
-v ~/.llama:/root/.llama \
-v $CUSTOM_RUN_CONFIG:/app/custom-run.yaml \
-e RUN_CONFIG_PATH=/app/custom-run.yaml \
llamastack/distribution-meta-reference-gpu \
--port $LLAMA_STACK_PORT
```
**Note**: The run configuration must be mounted into the container before it can be used. The `-v` flag mounts your local file into the container, and the `RUN_CONFIG_PATH` environment variable tells the entrypoint script which configuration to use.
Available run configurations for this distribution:
- `run.yaml`
- `run-with-safety.yaml`
### Via venv ### Via venv
Make sure you have the Llama Stack CLI available. Make sure you have the Llama Stack CLI available.

View file

@ -127,39 +127,13 @@ docker run \
-it \ -it \
--pull always \ --pull always \
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
-v ~/.llama:/root/.llama \ -v ./run.yaml:/root/my-run.yaml \
-e NVIDIA_API_KEY=$NVIDIA_API_KEY \ -e NVIDIA_API_KEY=$NVIDIA_API_KEY \
llamastack/distribution-nvidia \ llamastack/distribution-nvidia \
--config /root/my-run.yaml \
--port $LLAMA_STACK_PORT --port $LLAMA_STACK_PORT
``` ```
### Via Docker with Custom Run Configuration
You can also run the Docker container with a custom run configuration file by mounting it into the container:
```bash
# Set the path to your custom run.yaml file
CUSTOM_RUN_CONFIG=/path/to/your/custom-run.yaml
LLAMA_STACK_PORT=8321
docker run \
-it \
--pull always \
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
-v ~/.llama:/root/.llama \
-v $CUSTOM_RUN_CONFIG:/app/custom-run.yaml \
-e RUN_CONFIG_PATH=/app/custom-run.yaml \
-e NVIDIA_API_KEY=$NVIDIA_API_KEY \
llamastack/distribution-nvidia \
--port $LLAMA_STACK_PORT
```
**Note**: The run configuration must be mounted into the container before it can be used. The `-v` flag mounts your local file into the container, and the `RUN_CONFIG_PATH` environment variable tells the entrypoint script which configuration to use.
Available run configurations for this distribution:
- `run.yaml`
- `run-with-safety.yaml`
### Via venv ### Via venv
If you've set up your local development environment, you can also install the distribution dependencies using your local virtual environment. If you've set up your local development environment, you can also install the distribution dependencies using your local virtual environment.

View file

@ -46,9 +46,6 @@ providers:
api_key: ${env.TOGETHER_API_KEY:=} api_key: ${env.TOGETHER_API_KEY:=}
- provider_id: bedrock - provider_id: bedrock
provider_type: remote::bedrock provider_type: remote::bedrock
config:
api_key: ${env.AWS_BEDROCK_API_KEY:=}
region_name: ${env.AWS_DEFAULT_REGION:=us-east-2}
- provider_id: ${env.NVIDIA_API_KEY:+nvidia} - provider_id: ${env.NVIDIA_API_KEY:+nvidia}
provider_type: remote::nvidia provider_type: remote::nvidia
config: config:
@ -259,9 +256,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: shields:
@ -289,5 +283,3 @@ vector_stores:
default_embedding_model: default_embedding_model:
provider_id: sentence-transformers provider_id: sentence-transformers
model_id: nomic-ai/nomic-embed-text-v1.5 model_id: nomic-ai/nomic-embed-text-v1.5
safety:
default_shield_id: llama-guard

View file

@ -46,9 +46,6 @@ providers:
api_key: ${env.TOGETHER_API_KEY:=} api_key: ${env.TOGETHER_API_KEY:=}
- provider_id: bedrock - provider_id: bedrock
provider_type: remote::bedrock provider_type: remote::bedrock
config:
api_key: ${env.AWS_BEDROCK_API_KEY:=}
region_name: ${env.AWS_DEFAULT_REGION:=us-east-2}
- provider_id: ${env.NVIDIA_API_KEY:+nvidia} - provider_id: ${env.NVIDIA_API_KEY:+nvidia}
provider_type: remote::nvidia provider_type: remote::nvidia
config: config:
@ -250,9 +247,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: shields:
@ -280,5 +274,3 @@ vector_stores:
default_embedding_model: default_embedding_model:
provider_id: sentence-transformers provider_id: sentence-transformers
model_id: nomic-ai/nomic-embed-text-v1.5 model_id: nomic-ai/nomic-embed-text-v1.5
safety:
default_shield_id: llama-guard

View file

@ -109,9 +109,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: models:
- metadata: {} - metadata: {}

View file

@ -105,9 +105,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: models:
- metadata: {} - metadata: {}

View file

@ -122,9 +122,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: models:
- metadata: {} - metadata: {}

View file

@ -112,9 +112,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: models:
- metadata: {} - metadata: {}

View file

@ -111,9 +111,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: models:
- metadata: {} - metadata: {}

View file

@ -100,9 +100,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: [] shields: []

View file

@ -142,9 +142,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: models:
- metadata: {} - metadata: {}

View file

@ -46,9 +46,6 @@ providers:
api_key: ${env.TOGETHER_API_KEY:=} api_key: ${env.TOGETHER_API_KEY:=}
- provider_id: bedrock - provider_id: bedrock
provider_type: remote::bedrock provider_type: remote::bedrock
config:
api_key: ${env.AWS_BEDROCK_API_KEY:=}
region_name: ${env.AWS_DEFAULT_REGION:=us-east-2}
- provider_id: ${env.NVIDIA_API_KEY:+nvidia} - provider_id: ${env.NVIDIA_API_KEY:+nvidia}
provider_type: remote::nvidia provider_type: remote::nvidia
config: config:
@ -262,9 +259,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: shields:
@ -292,5 +286,3 @@ vector_stores:
default_embedding_model: default_embedding_model:
provider_id: sentence-transformers provider_id: sentence-transformers
model_id: nomic-ai/nomic-embed-text-v1.5 model_id: nomic-ai/nomic-embed-text-v1.5
safety:
default_shield_id: llama-guard

View file

@ -46,9 +46,6 @@ providers:
api_key: ${env.TOGETHER_API_KEY:=} api_key: ${env.TOGETHER_API_KEY:=}
- provider_id: bedrock - provider_id: bedrock
provider_type: remote::bedrock provider_type: remote::bedrock
config:
api_key: ${env.AWS_BEDROCK_API_KEY:=}
region_name: ${env.AWS_DEFAULT_REGION:=us-east-2}
- provider_id: ${env.NVIDIA_API_KEY:+nvidia} - provider_id: ${env.NVIDIA_API_KEY:+nvidia}
provider_type: remote::nvidia provider_type: remote::nvidia
config: config:
@ -253,9 +250,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: shields:
@ -283,5 +277,3 @@ vector_stores:
default_embedding_model: default_embedding_model:
provider_id: sentence-transformers provider_id: sentence-transformers
model_id: nomic-ai/nomic-embed-text-v1.5 model_id: nomic-ai/nomic-embed-text-v1.5
safety:
default_shield_id: llama-guard

View file

@ -46,9 +46,6 @@ providers:
api_key: ${env.TOGETHER_API_KEY:=} api_key: ${env.TOGETHER_API_KEY:=}
- provider_id: bedrock - provider_id: bedrock
provider_type: remote::bedrock provider_type: remote::bedrock
config:
api_key: ${env.AWS_BEDROCK_API_KEY:=}
region_name: ${env.AWS_DEFAULT_REGION:=us-east-2}
- provider_id: ${env.NVIDIA_API_KEY:+nvidia} - provider_id: ${env.NVIDIA_API_KEY:+nvidia}
provider_type: remote::nvidia provider_type: remote::nvidia
config: config:
@ -259,9 +256,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: shields:
@ -289,5 +283,3 @@ vector_stores:
default_embedding_model: default_embedding_model:
provider_id: sentence-transformers provider_id: sentence-transformers
model_id: nomic-ai/nomic-embed-text-v1.5 model_id: nomic-ai/nomic-embed-text-v1.5
safety:
default_shield_id: llama-guard

View file

@ -46,9 +46,6 @@ providers:
api_key: ${env.TOGETHER_API_KEY:=} api_key: ${env.TOGETHER_API_KEY:=}
- provider_id: bedrock - provider_id: bedrock
provider_type: remote::bedrock provider_type: remote::bedrock
config:
api_key: ${env.AWS_BEDROCK_API_KEY:=}
region_name: ${env.AWS_DEFAULT_REGION:=us-east-2}
- provider_id: ${env.NVIDIA_API_KEY:+nvidia} - provider_id: ${env.NVIDIA_API_KEY:+nvidia}
provider_type: remote::nvidia provider_type: remote::nvidia
config: config:
@ -250,9 +247,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: shields:
@ -280,5 +274,3 @@ vector_stores:
default_embedding_model: default_embedding_model:
provider_id: sentence-transformers provider_id: sentence-transformers
model_id: nomic-ai/nomic-embed-text-v1.5 model_id: nomic-ai/nomic-embed-text-v1.5
safety:
default_shield_id: llama-guard

View file

@ -35,6 +35,7 @@ from llama_stack.providers.remote.vector_io.pgvector.config import (
) )
from llama_stack.providers.remote.vector_io.qdrant.config import QdrantVectorIOConfig from llama_stack.providers.remote.vector_io.qdrant.config import QdrantVectorIOConfig
from llama_stack.providers.remote.vector_io.weaviate.config import WeaviateVectorIOConfig from llama_stack.providers.remote.vector_io.weaviate.config import WeaviateVectorIOConfig
from llama_stack.providers.utils.kvstore.config import PostgresKVStoreConfig
from llama_stack.providers.utils.sqlstore.sqlstore import PostgresSqlStoreConfig from llama_stack.providers.utils.sqlstore.sqlstore import PostgresSqlStoreConfig
@ -251,9 +252,6 @@ def get_distribution_template(name: str = "starter") -> DistributionTemplate:
model_id="nomic-ai/nomic-embed-text-v1.5", model_id="nomic-ai/nomic-embed-text-v1.5",
), ),
), ),
safety_config=SafetyConfig(
default_shield_id="llama-guard",
),
) )
postgres_run_settings = base_run_settings.model_copy( postgres_run_settings = base_run_settings.model_copy(

View file

@ -115,9 +115,6 @@ storage:
conversations: conversations:
table_name: openai_conversations table_name: openai_conversations
backend: sql_default backend: sql_default
prompts:
namespace: prompts
backend: kv_default
registered_resources: registered_resources:
models: [] models: []
shields: [] shields: []

1133
uv.lock generated

File diff suppressed because it is too large Load diff