From ca1b5041a486d7eed964bbf0a31946e329d45ca7 Mon Sep 17 00:00:00 2001 From: Mustafa Elbehery Date: Thu, 10 Jul 2025 15:41:27 +0200 Subject: [PATCH] chore: add mypy sentence transformers Signed-off-by: Mustafa Elbehery --- .../sentence_transformers/sentence_transformers.py | 6 +++--- pyproject.toml | 1 - 2 files changed, 3 insertions(+), 4 deletions(-) diff --git a/llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers.py b/llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers.py index 890c526f5..c6c419c41 100644 --- a/llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers.py +++ b/llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers.py @@ -34,7 +34,7 @@ from .config import SentenceTransformersInferenceConfig log = logging.getLogger(__name__) -class SentenceTransformersInferenceImpl( +class SentenceTransformersInferenceImpl( # type: ignore[misc] # Multiple inheritance from mixins creates MRO conflicts since sentence transformers don't actually support chat/completion operations OpenAIChatCompletionToLlamaStackMixin, OpenAICompletionToLlamaStackMixin, SentenceTransformerEmbeddingMixin, @@ -59,7 +59,7 @@ class SentenceTransformersInferenceImpl( async def completion( self, model_id: str, - content: str, + content: InterleavedContent, sampling_params: SamplingParams | None = None, response_format: ResponseFormat | None = None, stream: bool | None = False, @@ -72,10 +72,10 @@ class SentenceTransformersInferenceImpl( model_id: str, messages: list[Message], sampling_params: SamplingParams | None = None, - response_format: ResponseFormat | None = None, tools: list[ToolDefinition] | None = None, tool_choice: ToolChoice | None = ToolChoice.auto, tool_prompt_format: ToolPromptFormat | None = None, + response_format: ResponseFormat | None = None, stream: bool | None = False, logprobs: LogProbConfig | None = None, tool_config: ToolConfig | None = None, diff --git a/pyproject.toml b/pyproject.toml index 30598e5e3..555aa5f19 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -253,7 +253,6 @@ exclude = [ "^llama_stack/models/llama/llama4/", "^llama_stack/providers/inline/inference/meta_reference/parallel_utils\\.py$", "^llama_stack/providers/inline/inference/meta_reference/quantization/fp8_impls\\.py$", - "^llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers\\.py$", "^llama_stack/providers/inline/inference/vllm/", "^llama_stack/providers/inline/post_training/common/validator\\.py$", "^llama_stack/providers/inline/safety/code_scanner/",