From 7e5020b61f5a30066960da6378fd605ce9087767 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Mon, 27 Oct 2025 23:14:49 -0700 Subject: [PATCH] fix(mypy): use correct OpenAIChatCompletionChunk import in vllm MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Import OpenAIChatCompletionChunk from llama_stack.apis.inference instead of aliasing from openai package to match parent class signature. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude --- src/llama_stack/providers/remote/inference/vllm/vllm.py | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/src/llama_stack/providers/remote/inference/vllm/vllm.py b/src/llama_stack/providers/remote/inference/vllm/vllm.py index 74a18f3de..fa350ec48 100644 --- a/src/llama_stack/providers/remote/inference/vllm/vllm.py +++ b/src/llama_stack/providers/remote/inference/vllm/vllm.py @@ -7,13 +7,11 @@ from collections.abc import AsyncIterator from urllib.parse import urljoin import httpx -from openai.types.chat.chat_completion_chunk import ( - ChatCompletionChunk as OpenAIChatCompletionChunk, -) from pydantic import ConfigDict from llama_stack.apis.inference import ( OpenAIChatCompletion, + OpenAIChatCompletionChunk, OpenAIChatCompletionRequestWithExtraBody, ToolChoice, )