llama-stack-mirror/llama_stack/apis/agents
Shabana Baig add64e8e2a
feat: Add instructions parameter in response object (#3741)
# Problem
The current inline provider appends the user provided instructions to
messages as a system prompt, but the returned response object does not
contain the instructions field (as specified in the OpenAI responses
spec).

# What does this PR do?
This pull request adds the instruction field to the response object
definition and updates the inline provider. It also ensures that
instructions from previous response is not carried over to the next
response (as specified in the openAI spec).

Closes #[3566](https://github.com/llamastack/llama-stack/issues/3566)

## Test Plan

- Tested manually for change in model response w.r.t supplied
instructions field.
- Added unit test to check that the instructions from previous response
is not carried over to the next response.
- Added integration tests to check instructions parameter in the
returned response object.
- Added new recordings for the integration tests.

---------

Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-10-20 13:10:37 -07:00
..
__init__.py chore: remove nested imports (#2515) 2025-06-26 08:01:05 +05:30
agents.py feat: Add responses and safety impl extra_body (#3781) 2025-10-15 15:01:37 -07:00
openai_responses.py feat: Add instructions parameter in response object (#3741) 2025-10-20 13:10:37 -07:00