llama-stack-mirror/tests/verifications/openai_api
Ben Browning 8bee2954be
feat: Structured output for Responses API (#2324)
# What does this PR do?

This adds the missing `text` parameter to the Responses API that is how
users control structured outputs. All we do with that parameter is map
it to the corresponding chat completion response_format.

## Test Plan

The new unit tests exercise the various permutations allowed for this
property, while a couple of new verification tests actually use it for
real to verify the model outputs are following the format as expected.

Unit tests:

`python -m pytest -s -v
tests/unit/providers/agents/meta_reference/test_openai_responses.py`

Verification tests:

```
llama stack run llama_stack/templates/together/run.yaml
pytest -s -vv 'tests/verifications/openai_api/test_responses.py' \
  --base-url=http://localhost:8321/v1/openai/v1 \
  --model meta-llama/Llama-4-Scout-17B-16E-Instruct
```

Note that the verification tests can only be run with a real Llama Stack
server (as opposed to using the library client via
`--provider=stack:together`) because the Llama Stack python client is
not yet updated to accept this text field.

Signed-off-by: Ben Browning <bbrownin@redhat.com>
2025-06-03 14:43:00 -07:00
..
fixtures feat(responses): implement full multi-turn support (#2295) 2025-06-02 15:35:49 -07:00
__init__.py feat(verification): various improvements (#1921) 2025-04-10 10:26:19 -07:00
conftest.py feat(responses): add output_text delta events to responses (#2265) 2025-05-27 13:07:14 -07:00
test_chat_completion.py feat: OpenAI Responses API (#1989) 2025-04-28 14:06:00 -07:00
test_responses.py feat: Structured output for Responses API (#2324) 2025-06-03 14:43:00 -07:00