formatting issues

This commit is contained in:
Omar Abdelwahab 2025-11-06 12:43:24 -08:00
parent 5ce48d2c6a
commit d08c529ac0

View file

@ -24,12 +24,7 @@ from .fixtures.test_cases import (
multi_turn_tool_execution_test_cases, multi_turn_tool_execution_test_cases,
web_search_test_cases, web_search_test_cases,
) )
from .helpers import ( from .helpers import new_vector_store, setup_mcp_tools, upload_file, wait_for_file_attachment
new_vector_store,
setup_mcp_tools,
upload_file,
wait_for_file_attachment,
)
from .streaming_assertions import StreamingValidator from .streaming_assertions import StreamingValidator
@ -53,12 +48,7 @@ def test_response_non_streaming_web_search(compat_client, text_model_id, case):
@pytest.mark.parametrize("case", file_search_test_cases) @pytest.mark.parametrize("case", file_search_test_cases)
def test_response_non_streaming_file_search( def test_response_non_streaming_file_search(
compat_client, compat_client, text_model_id, embedding_model_id, embedding_dimension, tmp_path, case
text_model_id,
embedding_model_id,
embedding_dimension,
tmp_path,
case,
): ):
if isinstance(compat_client, LlamaStackAsLibraryClient): if isinstance(compat_client, LlamaStackAsLibraryClient):
pytest.skip("Responses API file search is not yet supported in library client.") pytest.skip("Responses API file search is not yet supported in library client.")
@ -258,8 +248,7 @@ def test_response_non_streaming_mcp_tool(compat_client, text_model_id, case, cap
) )
# Suppress expected auth error logs only for the failing auth attempt # Suppress expected auth error logs only for the failing auth attempt
with caplog.at_level( with caplog.at_level(
logging.CRITICAL, logging.CRITICAL, logger="llama_stack.providers.inline.agents.meta_reference.responses.streaming"
logger="llama_stack.providers.inline.agents.meta_reference.responses.streaming",
): ):
with pytest.raises(exc_type): with pytest.raises(exc_type):
compat_client.responses.create( compat_client.responses.create(
@ -323,11 +312,7 @@ def test_response_sequential_mcp_tool(compat_client, text_model_id, case):
assert "boiling point" in text_content.lower() assert "boiling point" in text_content.lower()
response2 = compat_client.responses.create( response2 = compat_client.responses.create(
model=text_model_id, model=text_model_id, input=case.input, tools=tools, stream=False, previous_response_id=response.id
input=case.input,
tools=tools,
stream=False,
previous_response_id=response.id,
) )
assert len(response2.output) >= 1 assert len(response2.output) >= 1
@ -376,13 +361,7 @@ def test_response_mcp_tool_approval(compat_client, text_model_id, case, approve)
response = compat_client.responses.create( response = compat_client.responses.create(
previous_response_id=response.id, previous_response_id=response.id,
model=text_model_id, model=text_model_id,
input=[ input=[{"type": "mcp_approval_response", "approval_request_id": approval_request.id, "approve": approve}],
{
"type": "mcp_approval_response",
"approval_request_id": approval_request.id,
"approve": approve,
}
],
tools=tools, tools=tools,
stream=False, stream=False,
) )
@ -459,11 +438,7 @@ def test_response_function_call_ordering_1(compat_client, text_model_id, case):
} }
) )
response = compat_client.responses.create( response = compat_client.responses.create(
model=text_model_id, model=text_model_id, input=inputs, tools=case.tools, stream=False, previous_response_id=response.id
input=inputs,
tools=case.tools,
stream=False,
previous_response_id=response.id,
) )
assert len(response.output) == 1 assert len(response.output) == 1