From a66b4c4c81eb2ea899cec0cd8cf3e1401b5c1b51 Mon Sep 17 00:00:00 2001 From: Yuan Tang Date: Tue, 18 Feb 2025 23:52:15 -0500 Subject: [PATCH] test: Enable test_text_chat_completion_with_tool_choice_required for remote::vllm (#1148) --- tests/client-sdk/inference/test_text_inference.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/tests/client-sdk/inference/test_text_inference.py b/tests/client-sdk/inference/test_text_inference.py index 52d5a24f2..6a113c463 100644 --- a/tests/client-sdk/inference/test_text_inference.py +++ b/tests/client-sdk/inference/test_text_inference.py @@ -250,8 +250,6 @@ def test_text_chat_completion_with_tool_calling_and_streaming( def test_text_chat_completion_with_tool_choice_required( llama_stack_client, text_model_id, get_weather_tool_definition, provider_tool_format, inference_provider_type ): - if inference_provider_type == "remote::vllm": - pytest.xfail("vllm-project/vllm#13002") response = llama_stack_client.inference.chat_completion( model_id=text_model_id, messages=[