{ "created": 1744154576.251519, "duration": 51.50739002227783, "exitcode": 0, "root": "/Users/erichuang/projects/llama-stack", "environment": {}, "summary": { "skipped": 61, "passed": 22, "total": 83, "collected": 83 }, "collectors": [ { "nodeid": "", "outcome": "passed", "result": [ { "nodeid": "tests/verifications/openai/test_chat_completion.py", "type": "Module" } ] }, { "nodeid": "tests/verifications/openai/test_chat_completion.py", "outcome": "passed", "result": [ { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o-mini]", "type": "Function", "lineno": 25 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o-mini]", "type": "Function", "lineno": 40 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 60 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 60 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 60 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 60 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o]", "type": "Function", "lineno": 60 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 60 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 75 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 75 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 75 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 75 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o]", "type": "Function", "lineno": 75 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 75 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]", "type": "Function", "lineno": 95 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o-mini]", "type": "Function", "lineno": 117 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]", "type": "Function", "lineno": 138 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]", "type": "Function", "lineno": 138 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]", "type": "Function", "lineno": 138 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]", "type": "Function", "lineno": 138 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "type": "Function", "lineno": 138 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o]", "type": "Function", "lineno": 138 }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]", "type": "Function", "lineno": 138 } ] } ], "tests": [ { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output0-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.0531630830373615, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.0001657919492572546, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output0-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006063499953597784, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.00014004099648445845, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.005356832989491522, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00016508297994732857, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006139832898043096, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00014450005255639553, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.00542324990965426, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00014112505596131086, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.004965625004842877, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.00013720791321247816, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o]", "lineno": 25, "outcome": "passed", "keywords": [ "test_chat_non_streaming_basic[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.005054084002040327, "outcome": "passed" }, "call": { "duration": 0.6271341659594327, "outcome": "passed" }, "teardown": { "duration": 0.00043925002682954073, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output0-gpt-4o-mini]", "lineno": 25, "outcome": "passed", "keywords": [ "test_chat_non_streaming_basic[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.0159178749890998, "outcome": "passed" }, "call": { "duration": 0.44088316697161645, "outcome": "passed" }, "teardown": { "duration": 0.0006467089988291264, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output1-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.016705541987903416, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.0005769169656559825, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output1-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.012067249976098537, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.00016683305148035288, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009295083000324667, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00017204193864017725, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009534333017654717, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00020175008103251457, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006628665956668556, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.0003687090938910842, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 25, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.0061322919791564345, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 26, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.0003664169926196337, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o]", "lineno": 25, "outcome": "passed", "keywords": [ "test_chat_non_streaming_basic[input_output1-gpt-4o]", "parametrize", "pytestmark", "input_output1-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.00623433303553611, "outcome": "passed" }, "call": { "duration": 0.7898445830214769, "outcome": "passed" }, "teardown": { "duration": 0.0006602079374715686, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_basic[input_output1-gpt-4o-mini]", "lineno": 25, "outcome": "passed", "keywords": [ "test_chat_non_streaming_basic[input_output1-gpt-4o-mini]", "parametrize", "pytestmark", "input_output1-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.014758958015590906, "outcome": "passed" }, "call": { "duration": 1.1555478329537436, "outcome": "passed" }, "teardown": { "duration": 0.0011781250359490514, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output0-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.03454475000035018, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.000967124942690134, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output0-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.025206666090525687, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.000189624959602952, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.014331333106383681, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00023133307695388794, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009339665994048119, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00020329200197011232, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.010387042071670294, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00018254201859235764, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.012297999928705394, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.00018662505317479372, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o]", "lineno": 40, "outcome": "passed", "keywords": [ "test_chat_streaming_basic[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006984042003750801, "outcome": "passed" }, "call": { "duration": 0.32529433304443955, "outcome": "passed" }, "teardown": { "duration": 0.0033042499562725425, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output0-gpt-4o-mini]", "lineno": 40, "outcome": "passed", "keywords": [ "test_chat_streaming_basic[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.01832079200539738, "outcome": "passed" }, "call": { "duration": 0.48440287495031953, "outcome": "passed" }, "teardown": { "duration": 0.00047233293298631907, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output1-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.02893691696226597, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.0001747499918565154, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output1-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006553041050210595, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.00016829196829348803, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.013746666954830289, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00019237503875046968, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output1-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007175332983024418, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.0001873329747468233, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006127291941083968, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00019004102796316147, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 40, "outcome": "skipped", "keywords": [ "test_chat_streaming_basic[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006421791040338576, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 41, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.0001611249754205346, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o]", "lineno": 40, "outcome": "passed", "keywords": [ "test_chat_streaming_basic[input_output1-gpt-4o]", "parametrize", "pytestmark", "input_output1-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009806249989196658, "outcome": "passed" }, "call": { "duration": 0.9556747920578346, "outcome": "passed" }, "teardown": { "duration": 0.0004937920020893216, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_basic[input_output1-gpt-4o-mini]", "lineno": 40, "outcome": "passed", "keywords": [ "test_chat_streaming_basic[input_output1-gpt-4o-mini]", "parametrize", "pytestmark", "input_output1-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.03146500000730157, "outcome": "passed" }, "call": { "duration": 1.082494750036858, "outcome": "passed" }, "teardown": { "duration": 0.0006242080125957727, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]", "lineno": 60, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.021534667001105845, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.0003469999646767974, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 60, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.025929750059731305, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.0008774169255048037, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 60, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.012507125036790967, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00022008304949849844, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 60, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.008156375028192997, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 61, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.0002079169498756528, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o]", "lineno": 60, "outcome": "passed", "keywords": [ "test_chat_non_streaming_image[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.012587749981321394, "outcome": "passed" }, "call": { "duration": 2.7379885419504717, "outcome": "passed" }, "teardown": { "duration": 0.00044579198583960533, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_image[input_output0-gpt-4o-mini]", "lineno": 60, "outcome": "passed", "keywords": [ "test_chat_non_streaming_image[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.017111250082962215, "outcome": "passed" }, "call": { "duration": 2.599374584038742, "outcome": "passed" }, "teardown": { "duration": 0.0009177909232676029, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]", "lineno": 75, "outcome": "skipped", "keywords": [ "test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.02198700001463294, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00042749999556690454, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 75, "outcome": "skipped", "keywords": [ "test_chat_streaming_image[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.015032917028293014, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00041016703471541405, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 75, "outcome": "skipped", "keywords": [ "test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.013976250076666474, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00027600000612437725, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 75, "outcome": "skipped", "keywords": [ "test_chat_streaming_image[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.00799729092977941, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 76, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.00020320899784564972, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o]", "lineno": 75, "outcome": "passed", "keywords": [ "test_chat_streaming_image[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.010483540943823755, "outcome": "passed" }, "call": { "duration": 4.249965250026435, "outcome": "passed" }, "teardown": { "duration": 0.0008596250554546714, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_image[input_output0-gpt-4o-mini]", "lineno": 75, "outcome": "passed", "keywords": [ "test_chat_streaming_image[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.018141582957468927, "outcome": "passed" }, "call": { "duration": 2.297856790944934, "outcome": "passed" }, "teardown": { "duration": 0.0005075830267742276, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.017144332989118993, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.0006829580524936318, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009827250032685697, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.00024204188957810402, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006737958989106119, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00022729102056473494, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006030917051248252, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00022229203023016453, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009183833957649767, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00022629194427281618, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007097500027157366, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.00826825003605336, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o]", "lineno": 95, "outcome": "passed", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006604874972254038, "outcome": "passed" }, "call": { "duration": 1.4057738750707358, "outcome": "passed" }, "teardown": { "duration": 0.000506040989421308, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]", "lineno": 95, "outcome": "passed", "keywords": [ "test_chat_non_streaming_structured_output[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.015966624952852726, "outcome": "passed" }, "call": { "duration": 0.540478374925442, "outcome": "passed" }, "teardown": { "duration": 0.0009536249563097954, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.020631707971915603, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.0004928340204060078, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.016745459055528045, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.0003412909572944045, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.012252667103894055, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00028650008607655764, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.01128904102370143, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00027041707653552294, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.009191332967020571, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.0002074999501928687, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 95, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007687666919082403, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 96, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.0002027079463005066, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o]", "lineno": 95, "outcome": "passed", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-gpt-4o]", "parametrize", "pytestmark", "input_output1-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007542708073742688, "outcome": "passed" }, "call": { "duration": 4.244797708000988, "outcome": "passed" }, "teardown": { "duration": 0.0012778330128639936, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]", "lineno": 95, "outcome": "passed", "keywords": [ "test_chat_non_streaming_structured_output[input_output1-gpt-4o-mini]", "parametrize", "pytestmark", "input_output1-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.026919999974779785, "outcome": "passed" }, "call": { "duration": 9.006108874920756, "outcome": "passed" }, "teardown": { "duration": 0.00046324997674673796, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output0-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.01554666692391038, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.0004023330984637141, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output0-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007354958914220333, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.0002900830004364252, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.017274250043556094, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.0002668329980224371, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006813667016103864, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.00024500000290572643, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007385291974060237, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.00017024995759129524, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.00857366609852761, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.00016850000247359276, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o]", "lineno": 117, "outcome": "passed", "keywords": [ "test_chat_streaming_structured_output[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.005570041947066784, "outcome": "passed" }, "call": { "duration": 0.8564215000951663, "outcome": "passed" }, "teardown": { "duration": 0.0004029169213026762, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output0-gpt-4o-mini]", "lineno": 117, "outcome": "passed", "keywords": [ "test_chat_streaming_structured_output[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.00786762498319149, "outcome": "passed" }, "call": { "duration": 0.6419672920601442, "outcome": "passed" }, "teardown": { "duration": 0.0005102079594507813, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output1-Llama-3.3-8B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-8B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.017147499951533973, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-8B-Instruct')" }, "teardown": { "duration": 0.00032350001856684685, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output1-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.01194737502373755, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.0005004579434171319, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.010250666993670166, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00022554199676960707, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output1-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007847042055800557, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.000283458037301898, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.008078000042587519, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.0001794169656932354, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 117, "outcome": "skipped", "keywords": [ "test_chat_streaming_structured_output[input_output1-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output1-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.007204750087112188, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 118, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.00017725001089274883, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o]", "lineno": 117, "outcome": "passed", "keywords": [ "test_chat_streaming_structured_output[input_output1-gpt-4o]", "parametrize", "pytestmark", "input_output1-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006797667010687292, "outcome": "passed" }, "call": { "duration": 5.411579457926564, "outcome": "passed" }, "teardown": { "duration": 0.001134666963480413, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_streaming_structured_output[input_output1-gpt-4o-mini]", "lineno": 117, "outcome": "passed", "keywords": [ "test_chat_streaming_structured_output[input_output1-gpt-4o-mini]", "parametrize", "pytestmark", "input_output1-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.025059624924324453, "outcome": "passed" }, "call": { "duration": 9.112342999898829, "outcome": "passed" }, "teardown": { "duration": 0.0009202499641105533, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]", "lineno": 138, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-Llama-3.3-70B-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-3.3-70B-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.024287916952744126, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-3.3-70B-Instruct')" }, "teardown": { "duration": 0.00015587499365210533, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]", "lineno": 138, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006531457998789847, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E')" }, "teardown": { "duration": 0.00014670798555016518, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]", "lineno": 138, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-Llama-4-Scout-17B-16E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Scout-17B-16E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.006190375075675547, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Scout-17B-16E-Instruct')" }, "teardown": { "duration": 0.0001603750279173255, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]", "lineno": 138, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.005670750048011541, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E')" }, "teardown": { "duration": 0.0001479999627918005, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "lineno": 138, "outcome": "skipped", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-Llama-4-Maverick-17B-128E-Instruct]", "parametrize", "pytestmark", "input_output0-Llama-4-Maverick-17B-128E-Instruct", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.005662833107635379, "outcome": "skipped", "longrepr": "('/Users/erichuang/projects/llama-stack/tests/verifications/openai/test_chat_completion.py', 139, 'Skipped: Provider openai does not support model Llama-4-Maverick-17B-128E-Instruct')" }, "teardown": { "duration": 0.0001480829669162631, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o]", "lineno": 138, "outcome": "passed", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-gpt-4o]", "parametrize", "pytestmark", "input_output0-gpt-4o", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.00573637499473989, "outcome": "passed" }, "call": { "duration": 0.6269576249178499, "outcome": "passed" }, "teardown": { "duration": 0.0010142088867723942, "outcome": "passed" } }, { "nodeid": "tests/verifications/openai/test_chat_completion.py::test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]", "lineno": 138, "outcome": "passed", "keywords": [ "test_chat_non_streaming_tool_calling[input_output0-gpt-4o-mini]", "parametrize", "pytestmark", "input_output0-gpt-4o-mini", "test_chat_completion.py", "openai", "verifications", "tests", "llama-stack", "" ], "setup": { "duration": 0.01623620803002268, "outcome": "passed" }, "call": { "duration": 0.7144521250156686, "outcome": "passed" }, "teardown": { "duration": 0.0011040839599445462, "outcome": "passed" } } ] }