mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
# What does this PR do? This adds the missing `text` parameter to the Responses API that is how users control structured outputs. All we do with that parameter is map it to the corresponding chat completion response_format. ## Test Plan The new unit tests exercise the various permutations allowed for this property, while a couple of new verification tests actually use it for real to verify the model outputs are following the format as expected. Unit tests: `python -m pytest -s -v tests/unit/providers/agents/meta_reference/test_openai_responses.py` Verification tests: ``` llama stack run llama_stack/templates/together/run.yaml pytest -s -vv 'tests/verifications/openai_api/test_responses.py' \ --base-url=http://localhost:8321/v1/openai/v1 \ --model meta-llama/Llama-4-Scout-17B-16E-Instruct ``` Note that the verification tests can only be run with a real Llama Stack server (as opposed to using the library client via `--provider=stack:together`) because the Llama Stack python client is not yet updated to accept this text field. Signed-off-by: Ben Browning <bbrownin@redhat.com> |
||
|---|---|---|
| .. | ||
| agents | ||
| batch_inference | ||
| benchmarks | ||
| common | ||
| datasetio | ||
| datasets | ||
| eval | ||
| files | ||
| inference | ||
| inspect | ||
| models | ||
| post_training | ||
| providers | ||
| safety | ||
| scoring | ||
| scoring_functions | ||
| shields | ||
| synthetic_data_generation | ||
| telemetry | ||
| tools | ||
| vector_dbs | ||
| vector_io | ||
| __init__.py | ||
| datatypes.py | ||
| resource.py | ||
| version.py | ||