version: 2 distribution_spec: description: Use NVIDIA NIM for running LLM inference, evaluation and safety providers: inference: - remote::nvidia vector_io: - inline::faiss safety: - remote::nvidia agents: - inline::meta-reference telemetry: - inline::meta-reference eval: - remote::nvidia post_training: - remote::nvidia datasetio: - inline::localfs - remote::nvidia scoring: - inline::basic tool_runtime: - inline::rag-runtime image_type: conda additional_pip_packages: - aiosqlite - sqlalchemy[asyncio]