Distributions updates (slight updates to ollama, add inline-vllm and remote-vllm) (#408)

* remote vllm distro

* add inline-vllm details, fix things

* Write some docs
This commit is contained in:
Ashwin Bharambe 2024-11-08 18:09:39 -08:00 committed by GitHub
parent ba82021d4b
commit 4986e46188
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
19 changed files with 365 additions and 46 deletions

View file

@ -0,0 +1,13 @@
name: meta-reference-gpu
distribution_spec:
docker_image: pytorch/pytorch:2.5.0-cuda12.4-cudnn9-runtime
description: Use code from `llama_stack` itself to serve all llama stack APIs
providers:
inference: meta-reference
memory:
- meta-reference
- remote::chromadb
- remote::pgvector
safety: meta-reference
agents: meta-reference
telemetry: meta-reference