From f7d5e3f5d6126558ef6dbf8f3cddbef88c212ff4 Mon Sep 17 00:00:00 2001 From: Yuan Tang Date: Thu, 20 Mar 2025 20:44:59 -0400 Subject: [PATCH] test: Add default vLLM URL in remote-vllm template Signed-off-by: Yuan Tang --- llama_stack/templates/remote-vllm/run.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_stack/templates/remote-vllm/run.yaml b/llama_stack/templates/remote-vllm/run.yaml index b6bba1252..1f3cdfb39 100644 --- a/llama_stack/templates/remote-vllm/run.yaml +++ b/llama_stack/templates/remote-vllm/run.yaml @@ -15,7 +15,7 @@ providers: - provider_id: vllm-inference provider_type: remote::vllm config: - url: ${env.VLLM_URL} + url: ${env.VLLM_URL:http://localhost:8000/v1} max_tokens: ${env.VLLM_MAX_TOKENS:4096} api_token: ${env.VLLM_API_TOKEN:fake} tls_verify: ${env.VLLM_TLS_VERIFY:true}