name: local-ollama distribution_spec: description: Like local, but use ollama for running LLM inference providers: inference: remote::ollama memory: meta-reference safety: meta-reference agents: meta-reference telemetry: meta-reference image_type: conda