mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
When using bash style substitution env variable in distribution template, we are processing the string and convert it to the type associated with the provider's config class. This allows us to return the proper type. This is crucial for api key since they are not strings anymore but SecretStr. If the key is unset we will get an empty string which will result in a Pydantic error like: ``` ERROR 2025-09-25 21:40:44,565 __main__:527 core::server: Error creating app: 1 validation error for AnthropicConfig api_key Input should be a valid string For further information visit https://errors.pydantic.dev/2.11/v/string_type ``` Signed-off-by: Sébastien Han <seb@redhat.com>
41 lines
1.9 KiB
Python
41 lines
1.9 KiB
Python
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
from typing import Any
|
|
|
|
from pydantic import BaseModel, Field
|
|
|
|
from llama_stack.core.secret_types import MySecretStr
|
|
from llama_stack.providers.utils.sqlstore.sqlstore import SqliteSqlStoreConfig, SqlStoreConfig
|
|
|
|
|
|
class S3FilesImplConfig(BaseModel):
|
|
"""Configuration for S3-based files provider."""
|
|
|
|
bucket_name: str = Field(description="S3 bucket name to store files")
|
|
region: str = Field(default="us-east-1", description="AWS region where the bucket is located")
|
|
aws_access_key_id: str | None = Field(default=None, description="AWS access key ID (optional if using IAM roles)")
|
|
aws_secret_access_key: MySecretStr = Field(description="AWS secret access key (optional if using IAM roles)")
|
|
endpoint_url: str | None = Field(default=None, description="Custom S3 endpoint URL (for MinIO, LocalStack, etc.)")
|
|
auto_create_bucket: bool = Field(
|
|
default=False, description="Automatically create the S3 bucket if it doesn't exist"
|
|
)
|
|
metadata_store: SqlStoreConfig = Field(description="SQL store configuration for file metadata")
|
|
|
|
@classmethod
|
|
def sample_run_config(cls, __distro_dir__: str) -> dict[str, Any]:
|
|
return {
|
|
"bucket_name": "${env.S3_BUCKET_NAME}", # no default, buckets must be globally unique
|
|
"region": "${env.AWS_REGION:=us-east-1}",
|
|
"aws_access_key_id": "${env.AWS_ACCESS_KEY_ID:=}",
|
|
"aws_secret_access_key": "${env.AWS_SECRET_ACCESS_KEY:=}",
|
|
"endpoint_url": "${env.S3_ENDPOINT_URL:=}",
|
|
"auto_create_bucket": "${env.S3_AUTO_CREATE_BUCKET:=false}",
|
|
"metadata_store": SqliteSqlStoreConfig.sample_run_config(
|
|
__distro_dir__=__distro_dir__,
|
|
db_name="s3_files_metadata.db",
|
|
),
|
|
}
|