Add vLLM inference provider for OpenAI compatible vLLM server (#178)

This PR adds vLLM inference provider for OpenAI compatible vLLM server.
This commit is contained in:
Yuan Tang 2024-10-20 21:43:25 -04:00 committed by GitHub
parent 59c43736e8
commit a27a2cd2af
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
6 changed files with 209 additions and 1 deletions

View file

@ -60,6 +60,15 @@ def available_providers() -> List[ProviderSpec]:
module="llama_stack.providers.adapters.inference.ollama",
),
),
# remote_provider_spec(
# api=Api.inference,
# adapter=AdapterSpec(
# adapter_type="vllm",
# pip_packages=["openai"],
# module="llama_stack.providers.adapters.inference.vllm",
# config_class="llama_stack.providers.adapters.inference.vllm.VLLMImplConfig",
# ),
# ),
remote_provider_spec(
api=Api.inference,
adapter=AdapterSpec(