llama-stack-mirror/llama_stack/providers/inline/agents/meta_reference/responses
Ashwin Bharambe 30ba8c8655
fix(responses): sync conversation before yielding terminal events in streaming (#3888)
Move conversation sync logic before yield to ensure it executes even
when
streaming consumers break early after receiving response.completed
event.

## Test Plan

```
OLLAMA_URL=http://localhost:11434 \
  pytest -sv tests/integration/responses/ \
  --stack-config server:ci-tests \
  --text-model ollama/llama3.2:3b-instruct-fp16 \
  --inference-mode live \
  -k conversation_multi
```

This test now passes.
2025-10-22 14:31:12 -07:00
..
__init__.py chore(responses): Refactor Responses Impl to be civilized (#3138) 2025-08-15 00:05:35 +00:00
openai_responses.py fix(responses): sync conversation before yielding terminal events in streaming (#3888) 2025-10-22 14:31:12 -07:00
streaming.py feat: Add instructions parameter in response object (#3741) 2025-10-20 13:10:37 -07:00
tool_executor.py fix(responses): fix subtle bugs in non-function tool calling (#3817) 2025-10-15 13:57:37 -07:00
types.py feat: reuse previous mcp tool listings where possible (#3710) 2025-10-10 09:28:25 -07:00
utils.py feat: Add responses and safety impl extra_body (#3781) 2025-10-15 15:01:37 -07:00