name: remote-vllm distribution_spec: description: Use (an external) vLLM server for running LLM inference providers: inference: remote::vllm memory: - meta-reference - remote::chromadb - remote::pgvector safety: inline::llama-guard agents: meta-reference telemetry: meta-reference