From df2d833e72db86632cf057f509414f0a423963ca Mon Sep 17 00:00:00 2001 From: Yuan Tang Date: Fri, 28 Feb 2025 12:46:31 -0500 Subject: [PATCH] more concise Signed-off-by: Yuan Tang --- llama_stack/providers/remote/inference/vllm/vllm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_stack/providers/remote/inference/vllm/vllm.py b/llama_stack/providers/remote/inference/vllm/vllm.py index e80cdfb29..8ec23cd90 100644 --- a/llama_stack/providers/remote/inference/vllm/vllm.py +++ b/llama_stack/providers/remote/inference/vllm/vllm.py @@ -274,7 +274,7 @@ class VLLMInferenceAdapter(Inference, ModelsProtocolPrivate): # References: # * https://platform.openai.com/docs/api-reference/chat/create#chat-create-tool_choice # * https://github.com/vllm-project/vllm/pull/10000 - if (tools is None or len(tools) == 0) and tool_config is not None: + if not tools and tool_config is not None: tool_config.tool_choice = ToolChoice.none request = ChatCompletionRequest( model=model.provider_resource_id,