From 3fb643b847fa1f20c48675fe3171ba359063c679 Mon Sep 17 00:00:00 2001 From: Matthew Farrellee Date: Tue, 23 Sep 2025 10:46:07 -0400 Subject: [PATCH] fix vllm unit tests, openai.resources.models.AsyncModels.list is not async --- tests/unit/providers/inference/test_remote_vllm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/unit/providers/inference/test_remote_vllm.py b/tests/unit/providers/inference/test_remote_vllm.py index 9545e0cf6..4dc2e0c16 100644 --- a/tests/unit/providers/inference/test_remote_vllm.py +++ b/tests/unit/providers/inference/test_remote_vllm.py @@ -62,7 +62,7 @@ from llama_stack.providers.remote.inference.vllm.vllm import ( @pytest.fixture(scope="module") def mock_openai_models_list(): - with patch("openai.resources.models.AsyncModels.list", new_callable=AsyncMock) as mock_list: + with patch("openai.resources.models.AsyncModels.list") as mock_list: yield mock_list