From f573baeec04cc2175288a5245c44b81612f0a4af Mon Sep 17 00:00:00 2001 From: Mike Sager Date: Fri, 14 Nov 2025 14:56:24 -0500 Subject: [PATCH] fix: linter fixes --- .../meta_reference/test_openai_responses.py | 40 +++++++++---------- .../test_openai_responses_conversations.py | 14 +++---- .../test_response_conversion_utils.py | 40 +++++++++---------- .../test_response_tool_context.py | 2 +- .../test_responses_safety_utils.py | 4 +- 5 files changed, 50 insertions(+), 50 deletions(-) diff --git a/tests/unit/providers/agents/meta_reference/test_openai_responses.py b/tests/unit/providers/agents/meta_reference/test_openai_responses.py index 5f0cbab9c..78f0d7cfd 100644 --- a/tests/unit/providers/agents/meta_reference/test_openai_responses.py +++ b/tests/unit/providers/agents/meta_reference/test_openai_responses.py @@ -15,7 +15,27 @@ from openai.types.chat.chat_completion_chunk import ( ChoiceDeltaToolCallFunction, ) +from llama_stack.core.access_control.access_control import default_policy +from llama_stack.core.storage.datatypes import ResponsesStoreReference, SqliteSqlStoreConfig +from llama_stack.providers.inline.agents.meta_reference.responses.openai_responses import ( + OpenAIResponsesImpl, +) +from llama_stack.providers.utils.responses.responses_store import ( + ResponsesStore, + _OpenAIResponseObjectWithInputAndMessages, +) +from llama_stack.providers.utils.sqlstore.sqlstore import register_sqlstore_backends from llama_stack_api.agents import Order +from llama_stack_api.inference import ( + OpenAIAssistantMessageParam, + OpenAIChatCompletionContentPartTextParam, + OpenAIChatCompletionRequestWithExtraBody, + OpenAIDeveloperMessageParam, + OpenAIJSONSchema, + OpenAIResponseFormatJSONObject, + OpenAIResponseFormatJSONSchema, + OpenAIUserMessageParam, +) from llama_stack_api.openai_responses import ( ListOpenAIResponseInputItem, OpenAIResponseInputMessageContentText, @@ -31,27 +51,7 @@ from llama_stack_api.openai_responses import ( OpenAIResponseTextFormat, WebSearchToolTypes, ) -from llama_stack_api.inference import ( - OpenAIAssistantMessageParam, - OpenAIChatCompletionContentPartTextParam, - OpenAIChatCompletionRequestWithExtraBody, - OpenAIDeveloperMessageParam, - OpenAIJSONSchema, - OpenAIResponseFormatJSONObject, - OpenAIResponseFormatJSONSchema, - OpenAIUserMessageParam, -) from llama_stack_api.tools import ListToolDefsResponse, ToolDef, ToolGroups, ToolInvocationResult, ToolRuntime -from llama_stack.core.access_control.access_control import default_policy -from llama_stack.core.storage.datatypes import ResponsesStoreReference, SqliteSqlStoreConfig -from llama_stack.providers.inline.agents.meta_reference.responses.openai_responses import ( - OpenAIResponsesImpl, -) -from llama_stack.providers.utils.responses.responses_store import ( - ResponsesStore, - _OpenAIResponseObjectWithInputAndMessages, -) -from llama_stack.providers.utils.sqlstore.sqlstore import register_sqlstore_backends from tests.unit.providers.agents.meta_reference.fixtures import load_chat_completion_fixture diff --git a/tests/unit/providers/agents/meta_reference/test_openai_responses_conversations.py b/tests/unit/providers/agents/meta_reference/test_openai_responses_conversations.py index 598346bf9..fa1ddae78 100644 --- a/tests/unit/providers/agents/meta_reference/test_openai_responses_conversations.py +++ b/tests/unit/providers/agents/meta_reference/test_openai_responses_conversations.py @@ -7,13 +7,6 @@ import pytest -from llama_stack_api.openai_responses import ( - OpenAIResponseMessage, - OpenAIResponseObject, - OpenAIResponseObjectStreamResponseCompleted, - OpenAIResponseObjectStreamResponseOutputItemDone, - OpenAIResponseOutputMessageContentOutputText, -) from llama_stack_api.common.errors import ( ConversationNotFoundError, InvalidConversationIdError, @@ -21,6 +14,13 @@ from llama_stack_api.common.errors import ( from llama_stack_api.conversations import ( ConversationItemList, ) +from llama_stack_api.openai_responses import ( + OpenAIResponseMessage, + OpenAIResponseObject, + OpenAIResponseObjectStreamResponseCompleted, + OpenAIResponseObjectStreamResponseOutputItemDone, + OpenAIResponseOutputMessageContentOutputText, +) # Import existing fixtures from the main responses test file pytest_plugins = ["tests.unit.providers.agents.meta_reference.test_openai_responses"] diff --git a/tests/unit/providers/agents/meta_reference/test_response_conversion_utils.py b/tests/unit/providers/agents/meta_reference/test_response_conversion_utils.py index 50cd6ef1d..b7a437686 100644 --- a/tests/unit/providers/agents/meta_reference/test_response_conversion_utils.py +++ b/tests/unit/providers/agents/meta_reference/test_response_conversion_utils.py @@ -7,18 +7,14 @@ import pytest -from llama_stack_api.openai_responses import ( - OpenAIResponseAnnotationFileCitation, - OpenAIResponseInputFunctionToolCallOutput, - OpenAIResponseInputMessageContentImage, - OpenAIResponseInputMessageContentText, - OpenAIResponseInputToolFunction, - OpenAIResponseInputToolWebSearch, - OpenAIResponseMessage, - OpenAIResponseOutputMessageContentOutputText, - OpenAIResponseOutputMessageFunctionToolCall, - OpenAIResponseText, - OpenAIResponseTextFormat, +from llama_stack.providers.inline.agents.meta_reference.responses.utils import ( + _extract_citations_from_text, + convert_chat_choice_to_response_message, + convert_response_content_to_chat_content, + convert_response_input_to_chat_messages, + convert_response_text_to_chat_response_format, + get_message_type_by_role, + is_function_tool_call, ) from llama_stack_api.inference import ( OpenAIAssistantMessageParam, @@ -35,14 +31,18 @@ from llama_stack_api.inference import ( OpenAIToolMessageParam, OpenAIUserMessageParam, ) -from llama_stack.providers.inline.agents.meta_reference.responses.utils import ( - _extract_citations_from_text, - convert_chat_choice_to_response_message, - convert_response_content_to_chat_content, - convert_response_input_to_chat_messages, - convert_response_text_to_chat_response_format, - get_message_type_by_role, - is_function_tool_call, +from llama_stack_api.openai_responses import ( + OpenAIResponseAnnotationFileCitation, + OpenAIResponseInputFunctionToolCallOutput, + OpenAIResponseInputMessageContentImage, + OpenAIResponseInputMessageContentText, + OpenAIResponseInputToolFunction, + OpenAIResponseInputToolWebSearch, + OpenAIResponseMessage, + OpenAIResponseOutputMessageContentOutputText, + OpenAIResponseOutputMessageFunctionToolCall, + OpenAIResponseText, + OpenAIResponseTextFormat, ) diff --git a/tests/unit/providers/agents/meta_reference/test_response_tool_context.py b/tests/unit/providers/agents/meta_reference/test_response_tool_context.py index 8d5312ec8..4054debd5 100644 --- a/tests/unit/providers/agents/meta_reference/test_response_tool_context.py +++ b/tests/unit/providers/agents/meta_reference/test_response_tool_context.py @@ -5,6 +5,7 @@ # the root directory of this source tree. +from llama_stack.providers.inline.agents.meta_reference.responses.types import ToolContext from llama_stack_api.openai_responses import ( MCPListToolsTool, OpenAIResponseInputToolFileSearch, @@ -15,7 +16,6 @@ from llama_stack_api.openai_responses import ( OpenAIResponseOutputMessageMCPListTools, OpenAIResponseToolMCP, ) -from llama_stack.providers.inline.agents.meta_reference.responses.types import ToolContext class TestToolContext: diff --git a/tests/unit/providers/agents/meta_reference/test_responses_safety_utils.py b/tests/unit/providers/agents/meta_reference/test_responses_safety_utils.py index f3404f6cb..d4d1b872a 100644 --- a/tests/unit/providers/agents/meta_reference/test_responses_safety_utils.py +++ b/tests/unit/providers/agents/meta_reference/test_responses_safety_utils.py @@ -8,8 +8,6 @@ from unittest.mock import AsyncMock import pytest -from llama_stack_api.agents import ResponseGuardrailSpec -from llama_stack_api.safety import ModerationObject, ModerationObjectResults from llama_stack.providers.inline.agents.meta_reference.responses.openai_responses import ( OpenAIResponsesImpl, ) @@ -17,6 +15,8 @@ from llama_stack.providers.inline.agents.meta_reference.responses.utils import ( extract_guardrail_ids, run_guardrails, ) +from llama_stack_api.agents import ResponseGuardrailSpec +from llama_stack_api.safety import ModerationObject, ModerationObjectResults @pytest.fixture