version: '2' built_at: '2024-10-08T17:40:45.325529' image_name: local docker_image: null conda_env: local apis: - shields - agents - models - memory - memory_banks - inference - safety providers: inference: - provider_id: ollama0 provider_type: remote::ollama config: url: http://127.0.0.1:14343 safety: - provider_id: meta0 provider_type: meta-reference config: llama_guard_shield: model: Llama-Guard-3-1B excluded_categories: [] disable_input_check: false disable_output_check: false prompt_guard_shield: model: Prompt-Guard-86M memory: - provider_id: meta0 provider_type: meta-reference config: {} agents: - provider_id: meta0 provider_type: meta-reference config: persistence_store: namespace: null type: sqlite db_path: ~/.llama/runtime/kvstore.db telemetry: - provider_id: meta0 provider_type: meta-reference config: {}