forked from phoenix-oss/llama-stack-mirror
# What does this PR do? ## Test Plan pytest verifications/openai/test_chat_completion.py --provider together
2672 lines
91 KiB
JSON
2672 lines
91 KiB
JSON
{
|
|
"created": 1744154576.251519,
|
|
"duration": 51.50739002227783,
|
|
"exitcode": 0,
|
|
"root": "/Users/erichuang/projects/llama-stack",
|
|
"environment": {},
|
|
"summary": {
|
|
"skipped": 61,
|
|
"passed": 22,
|
|
"total": 83,
|
|
"collected": 83
|
|
},
|
|
"collectors": [
|
|
{
|
|
"nodeid": "",
|
|
"outcome": "passed",
|
|
"result": [
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py",
|
|
"type": "Module"
|
|
}
|
|
]
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py",
|
|
"outcome": "passed",
|
|
"result": [
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 25
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 40
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 60
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 60
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 60
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 60
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 60
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 60
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 75
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 75
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 75
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 75
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 75
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 75
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 95
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 117
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]",
|
|
"type": "Function",
|
|
"lineno": 138
|
|
}
|
|
]
|
|
}
|
|
],
|
|
"tests": [
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.0531630830373615,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001657919492572546,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006063499953597784,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00014004099648445845,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.005356832989491522,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00016508297994732857,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006139832898043096,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00014450005255639553,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.00542324990965426,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00014112505596131086,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.004965625004842877,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00013720791321247816,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o]",
|
|
"lineno": 25,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.005054084002040327,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.6271341659594327,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00043925002682954073,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o-mini]",
|
|
"lineno": 25,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.0159178749890998,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.44088316697161645,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0006467089988291264,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.016705541987903416,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0005769169656559825,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.012067249976098537,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00016683305148035288,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009295083000324667,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00017204193864017725,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009534333017654717,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00020175008103251457,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006628665956668556,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0003687090938910842,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 25,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.0061322919791564345,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0003664169926196337,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o]",
|
|
"lineno": 25,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.00623433303553611,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.7898445830214769,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0006602079374715686,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o-mini]",
|
|
"lineno": 25,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_basic[input_output1-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.014758958015590906,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 1.1555478329537436,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0011781250359490514,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.03454475000035018,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.000967124942690134,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.025206666090525687,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.000189624959602952,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.014331333106383681,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00023133307695388794,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009339665994048119,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00020329200197011232,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.010387042071670294,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00018254201859235764,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.012297999928705394,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00018662505317479372,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o]",
|
|
"lineno": 40,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006984042003750801,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.32529433304443955,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0033042499562725425,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o-mini]",
|
|
"lineno": 40,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.01832079200539738,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.48440287495031953,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00047233293298631907,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.02893691696226597,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001747499918565154,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006553041050210595,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00016829196829348803,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.013746666954830289,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00019237503875046968,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007175332983024418,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001873329747468233,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006127291941083968,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00019004102796316147,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 40,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006421791040338576,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001611249754205346,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o]",
|
|
"lineno": 40,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009806249989196658,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.9556747920578346,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0004937920020893216,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o-mini]",
|
|
"lineno": 40,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_basic[input_output1-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.03146500000730157,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 1.082494750036858,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0006242080125957727,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 60,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.021534667001105845,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0003469999646767974,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 60,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.025929750059731305,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0008774169255048037,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 60,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.012507125036790967,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00022008304949849844,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 60,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.008156375028192997,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0002079169498756528,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o]",
|
|
"lineno": 60,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_image[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.012587749981321394,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 2.7379885419504717,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00044579198583960533,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o-mini]",
|
|
"lineno": 60,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_image[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.017111250082962215,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 2.599374584038742,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0009177909232676029,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 75,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.02198700001463294,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00042749999556690454,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 75,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.015032917028293014,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00041016703471541405,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 75,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.013976250076666474,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00027600000612437725,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 75,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.00799729092977941,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00020320899784564972,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o]",
|
|
"lineno": 75,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_image[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.010483540943823755,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 4.249965250026435,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0008596250554546714,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o-mini]",
|
|
"lineno": 75,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_image[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.018141582957468927,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 2.297856790944934,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0005075830267742276,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.017144332989118993,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0006829580524936318,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009827250032685697,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00024204188957810402,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006737958989106119,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00022729102056473494,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006030917051248252,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00022229203023016453,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009183833957649767,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00022629194427281618,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007097500027157366,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00826825003605336,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o]",
|
|
"lineno": 95,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006604874972254038,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 1.4057738750707358,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.000506040989421308,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]",
|
|
"lineno": 95,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.015966624952852726,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.540478374925442,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0009536249563097954,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.020631707971915603,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0004928340204060078,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.016745459055528045,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0003412909572944045,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.012252667103894055,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00028650008607655764,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.01128904102370143,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00027041707653552294,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.009191332967020571,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0002074999501928687,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 95,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007687666919082403,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0002027079463005066,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o]",
|
|
"lineno": 95,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007542708073742688,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 4.244797708000988,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0012778330128639936,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]",
|
|
"lineno": 95,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.026919999974779785,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 9.006108874920756,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00046324997674673796,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.01554666692391038,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0004023330984637141,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007354958914220333,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0002900830004364252,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.017274250043556094,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0002668329980224371,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006813667016103864,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00024500000290572643,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007385291974060237,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00017024995759129524,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.00857366609852761,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00016850000247359276,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o]",
|
|
"lineno": 117,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.005570041947066784,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.8564215000951663,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0004029169213026762,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o-mini]",
|
|
"lineno": 117,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.00786762498319149,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.6419672920601442,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0005102079594507813,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-8B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.017147499951533973,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00032350001856684685,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.01194737502373755,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0005004579434171319,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.010250666993670166,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00022554199676960707,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007847042055800557,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.000283458037301898,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.008078000042587519,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001794169656932354,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 117,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.007204750087112188,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00017725001089274883,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o]",
|
|
"lineno": 117,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006797667010687292,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 5.411579457926564,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.001134666963480413,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o-mini]",
|
|
"lineno": 117,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_streaming_structured_output[input_output1-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output1-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.025059624924324453,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 9.112342999898829,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0009202499641105533,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]",
|
|
"lineno": 138,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-3.3-70B-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.024287916952744126,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00015587499365210533,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]",
|
|
"lineno": 138,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006531457998789847,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.00014670798555016518,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"lineno": 138,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Scout-17B-16E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.006190375075675547,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001603750279173255,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"lineno": 138,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.005670750048011541,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001479999627918005,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"lineno": 138,
|
|
"outcome": "skipped",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.005662833107635379,
|
|
"outcome": "skipped",
|
|
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0001480829669162631,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o]",
|
|
"lineno": 138,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-gpt-4o]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.00573637499473989,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.6269576249178499,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0010142088867723942,
|
|
"outcome": "passed"
|
|
}
|
|
},
|
|
{
|
|
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]",
|
|
"lineno": 138,
|
|
"outcome": "passed",
|
|
"keywords": [
|
|
"test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]",
|
|
"parametrize",
|
|
"pytestmark",
|
|
"input_output0-gpt-4o-mini",
|
|
"test_chat_completion.py",
|
|
"openai",
|
|
"verifications",
|
|
"tests",
|
|
"llama-stack",
|
|
""
|
|
],
|
|
"setup": {
|
|
"duration": 0.01623620803002268,
|
|
"outcome": "passed"
|
|
},
|
|
"call": {
|
|
"duration": 0.7144521250156686,
|
|
"outcome": "passed"
|
|
},
|
|
"teardown": {
|
|
"duration": 0.0011040839599445462,
|
|
"outcome": "passed"
|
|
}
|
|
}
|
|
]
|
|
}
|