llama-stack/tests/verifications/test_results/openai_1744154522.json
ehhuang bcbc56baa2
feat: adds test suite to verify provider's OAI compat endpoints (#1901)
# What does this PR do?


## Test Plan
pytest verifications/openai/test_chat_completion.py --provider together
2025-04-08 21:21:38 -07:00

2672 lines
91 KiB
JSON

{
"created": 1744154576.251519,
"duration": 51.50739002227783,
"exitcode": 0,
"root": "/Users/erichuang/projects/llama-stack",
"environment": {},
"summary": {
"skipped": 61,
"passed": 22,
"total": 83,
"collected": 83
},
"collectors": [
{
"nodeid": "",
"outcome": "passed",
"result": [
{
"nodeid": "tests/verifications/openai/test_chat_completion.py",
"type": "Module"
}
]
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py",
"outcome": "passed",
"result": [
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o-mini]",
"type": "Function",
"lineno": 25
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o-mini]",
"type": "Function",
"lineno": 40
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 60
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 60
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 60
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 60
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o]",
"type": "Function",
"lineno": 60
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 60
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 75
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 75
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 75
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 75
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o]",
"type": "Function",
"lineno": 75
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 75
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]",
"type": "Function",
"lineno": 95
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o-mini]",
"type": "Function",
"lineno": 117
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]",
"type": "Function",
"lineno": 138
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]",
"type": "Function",
"lineno": 138
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"type": "Function",
"lineno": 138
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]",
"type": "Function",
"lineno": 138
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"type": "Function",
"lineno": 138
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o]",
"type": "Function",
"lineno": 138
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]",
"type": "Function",
"lineno": 138
}
]
}
],
"tests": [
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.0531630830373615,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.0001657919492572546,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006063499953597784,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.00014004099648445845,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.005356832989491522,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00016508297994732857,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006139832898043096,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00014450005255639553,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.00542324990965426,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00014112505596131086,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.004965625004842877,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.00013720791321247816,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o]",
"lineno": 25,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_basic[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.005054084002040327,
"outcome": "passed"
},
"call": {
"duration": 0.6271341659594327,
"outcome": "passed"
},
"teardown": {
"duration": 0.00043925002682954073,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o-mini]",
"lineno": 25,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_basic[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.0159178749890998,
"outcome": "passed"
},
"call": {
"duration": 0.44088316697161645,
"outcome": "passed"
},
"teardown": {
"duration": 0.0006467089988291264,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.016705541987903416,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.0005769169656559825,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.012067249976098537,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.00016683305148035288,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009295083000324667,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00017204193864017725,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009534333017654717,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00020175008103251457,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006628665956668556,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.0003687090938910842,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 25,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.0061322919791564345,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.0003664169926196337,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o]",
"lineno": 25,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_basic[input_output1-gpt-4o]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.00623433303553611,
"outcome": "passed"
},
"call": {
"duration": 0.7898445830214769,
"outcome": "passed"
},
"teardown": {
"duration": 0.0006602079374715686,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o-mini]",
"lineno": 25,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_basic[input_output1-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.014758958015590906,
"outcome": "passed"
},
"call": {
"duration": 1.1555478329537436,
"outcome": "passed"
},
"teardown": {
"duration": 0.0011781250359490514,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.03454475000035018,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.000967124942690134,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.025206666090525687,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.000189624959602952,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.014331333106383681,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00023133307695388794,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009339665994048119,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00020329200197011232,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.010387042071670294,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00018254201859235764,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.012297999928705394,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.00018662505317479372,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o]",
"lineno": 40,
"outcome": "passed",
"keywords": [
"test_chat_streaming_basic[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006984042003750801,
"outcome": "passed"
},
"call": {
"duration": 0.32529433304443955,
"outcome": "passed"
},
"teardown": {
"duration": 0.0033042499562725425,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o-mini]",
"lineno": 40,
"outcome": "passed",
"keywords": [
"test_chat_streaming_basic[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.01832079200539738,
"outcome": "passed"
},
"call": {
"duration": 0.48440287495031953,
"outcome": "passed"
},
"teardown": {
"duration": 0.00047233293298631907,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.02893691696226597,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.0001747499918565154,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006553041050210595,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.00016829196829348803,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.013746666954830289,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00019237503875046968,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007175332983024418,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.0001873329747468233,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006127291941083968,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00019004102796316147,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 40,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006421791040338576,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.0001611249754205346,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o]",
"lineno": 40,
"outcome": "passed",
"keywords": [
"test_chat_streaming_basic[input_output1-gpt-4o]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009806249989196658,
"outcome": "passed"
},
"call": {
"duration": 0.9556747920578346,
"outcome": "passed"
},
"teardown": {
"duration": 0.0004937920020893216,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o-mini]",
"lineno": 40,
"outcome": "passed",
"keywords": [
"test_chat_streaming_basic[input_output1-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.03146500000730157,
"outcome": "passed"
},
"call": {
"duration": 1.082494750036858,
"outcome": "passed"
},
"teardown": {
"duration": 0.0006242080125957727,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 60,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.021534667001105845,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.0003469999646767974,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 60,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.025929750059731305,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.0008774169255048037,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 60,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.012507125036790967,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00022008304949849844,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 60,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.008156375028192997,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.0002079169498756528,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o]",
"lineno": 60,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_image[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.012587749981321394,
"outcome": "passed"
},
"call": {
"duration": 2.7379885419504717,
"outcome": "passed"
},
"teardown": {
"duration": 0.00044579198583960533,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o-mini]",
"lineno": 60,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_image[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.017111250082962215,
"outcome": "passed"
},
"call": {
"duration": 2.599374584038742,
"outcome": "passed"
},
"teardown": {
"duration": 0.0009177909232676029,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 75,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.02198700001463294,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00042749999556690454,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 75,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.015032917028293014,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00041016703471541405,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 75,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.013976250076666474,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00027600000612437725,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 75,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.00799729092977941,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.00020320899784564972,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o]",
"lineno": 75,
"outcome": "passed",
"keywords": [
"test_chat_streaming_image[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.010483540943823755,
"outcome": "passed"
},
"call": {
"duration": 4.249965250026435,
"outcome": "passed"
},
"teardown": {
"duration": 0.0008596250554546714,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o-mini]",
"lineno": 75,
"outcome": "passed",
"keywords": [
"test_chat_streaming_image[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.018141582957468927,
"outcome": "passed"
},
"call": {
"duration": 2.297856790944934,
"outcome": "passed"
},
"teardown": {
"duration": 0.0005075830267742276,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.017144332989118993,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.0006829580524936318,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009827250032685697,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.00024204188957810402,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006737958989106119,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00022729102056473494,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006030917051248252,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00022229203023016453,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009183833957649767,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00022629194427281618,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007097500027157366,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.00826825003605336,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o]",
"lineno": 95,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006604874972254038,
"outcome": "passed"
},
"call": {
"duration": 1.4057738750707358,
"outcome": "passed"
},
"teardown": {
"duration": 0.000506040989421308,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]",
"lineno": 95,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.015966624952852726,
"outcome": "passed"
},
"call": {
"duration": 0.540478374925442,
"outcome": "passed"
},
"teardown": {
"duration": 0.0009536249563097954,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.020631707971915603,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.0004928340204060078,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.016745459055528045,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.0003412909572944045,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.012252667103894055,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00028650008607655764,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.01128904102370143,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00027041707653552294,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.009191332967020571,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.0002074999501928687,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 95,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007687666919082403,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.0002027079463005066,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o]",
"lineno": 95,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-gpt-4o]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007542708073742688,
"outcome": "passed"
},
"call": {
"duration": 4.244797708000988,
"outcome": "passed"
},
"teardown": {
"duration": 0.0012778330128639936,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]",
"lineno": 95,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.026919999974779785,
"outcome": "passed"
},
"call": {
"duration": 9.006108874920756,
"outcome": "passed"
},
"teardown": {
"duration": 0.00046324997674673796,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.01554666692391038,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.0004023330984637141,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007354958914220333,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.0002900830004364252,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.017274250043556094,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.0002668329980224371,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006813667016103864,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.00024500000290572643,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007385291974060237,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.00017024995759129524,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.00857366609852761,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.00016850000247359276,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o]",
"lineno": 117,
"outcome": "passed",
"keywords": [
"test_chat_streaming_structured_output[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.005570041947066784,
"outcome": "passed"
},
"call": {
"duration": 0.8564215000951663,
"outcome": "passed"
},
"teardown": {
"duration": 0.0004029169213026762,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o-mini]",
"lineno": 117,
"outcome": "passed",
"keywords": [
"test_chat_streaming_structured_output[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.00786762498319149,
"outcome": "passed"
},
"call": {
"duration": 0.6419672920601442,
"outcome": "passed"
},
"teardown": {
"duration": 0.0005102079594507813,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-8B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.017147499951533973,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')"
},
"teardown": {
"duration": 0.00032350001856684685,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.01194737502373755,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.0005004579434171319,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.010250666993670166,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00022554199676960707,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007847042055800557,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.000283458037301898,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.008078000042587519,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.0001794169656932354,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 117,
"outcome": "skipped",
"keywords": [
"test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output1-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.007204750087112188,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.00017725001089274883,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o]",
"lineno": 117,
"outcome": "passed",
"keywords": [
"test_chat_streaming_structured_output[input_output1-gpt-4o]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006797667010687292,
"outcome": "passed"
},
"call": {
"duration": 5.411579457926564,
"outcome": "passed"
},
"teardown": {
"duration": 0.001134666963480413,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o-mini]",
"lineno": 117,
"outcome": "passed",
"keywords": [
"test_chat_streaming_structured_output[input_output1-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output1-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.025059624924324453,
"outcome": "passed"
},
"call": {
"duration": 9.112342999898829,
"outcome": "passed"
},
"teardown": {
"duration": 0.0009202499641105533,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]",
"lineno": 138,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-3.3-70B-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.024287916952744126,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')"
},
"teardown": {
"duration": 0.00015587499365210533,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]",
"lineno": 138,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006531457998789847,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')"
},
"teardown": {
"duration": 0.00014670798555016518,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"lineno": 138,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Scout-17B-16E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.006190375075675547,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')"
},
"teardown": {
"duration": 0.0001603750279173255,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]",
"lineno": 138,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.005670750048011541,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')"
},
"teardown": {
"duration": 0.0001479999627918005,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"lineno": 138,
"outcome": "skipped",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]",
"parametrize",
"pytestmark",
"input_output0-Llama-4-Maverick-17B-128E-Instruct",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.005662833107635379,
"outcome": "skipped",
"longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')"
},
"teardown": {
"duration": 0.0001480829669162631,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o]",
"lineno": 138,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-gpt-4o]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.00573637499473989,
"outcome": "passed"
},
"call": {
"duration": 0.6269576249178499,
"outcome": "passed"
},
"teardown": {
"duration": 0.0010142088867723942,
"outcome": "passed"
}
},
{
"nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]",
"lineno": 138,
"outcome": "passed",
"keywords": [
"test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]",
"parametrize",
"pytestmark",
"input_output0-gpt-4o-mini",
"test_chat_completion.py",
"openai",
"verifications",
"tests",
"llama-stack",
""
],
"setup": {
"duration": 0.01623620803002268,
"outcome": "passed"
},
"call": {
"duration": 0.7144521250156686,
"outcome": "passed"
},
"teardown": {
"duration": 0.0011040839599445462,
"outcome": "passed"
}
}
]
}