From 31a0c51dea796aee3089e346c148cf523296bc5f Mon Sep 17 00:00:00 2001 From: Russell Bryant Date: Sat, 28 Sep 2024 19:06:53 +0000 Subject: [PATCH] Add vllm to the inference registry Signed-off-by: Russell Bryant --- llama_stack/providers/registry/inference.py | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/llama_stack/providers/registry/inference.py b/llama_stack/providers/registry/inference.py index 6cd97fd73..107bf69b2 100644 --- a/llama_stack/providers/registry/inference.py +++ b/llama_stack/providers/registry/inference.py @@ -104,4 +104,13 @@ def available_providers() -> List[ProviderSpec]: config_class="llama_stack.providers.adapters.inference.bedrock.BedrockConfig", ), ), + InlineProviderSpec( + api=Api.inference, + provider_id="vllm", + pip_packages=[ + "vllm", + ], + module="llama_stack.providers.impls.vllm", + config_class="llama_stack.providers.impls.vllm.VLLMConfig", + ), ]