llama-stack-mirror/llama_stack/providers/inline/agents/meta_reference
Ashwin Bharambe 8638537d14
feat(responses): stream progress of tool calls (#3135)
# What does this PR do?
Enhances tool execution streaming by adding support for real-time progress events during tool calls. This implementation adds streaming events for MCP and web search tools, including in-progress, searching, completed, and failed states. 

The refactored `_execute_tool_call` method now returns an async iterator that yields streaming events throughout the tool execution lifecycle.

## Test Plan
Updated the integration test `test_response_streaming_multi_turn_tool_execution` to verify the presence and structure of new streaming events, including:
- Checking for MCP in-progress and completed events
- Verifying that progress events contain required fields (item_id, output_index, sequence_number)
- Ensuring completed events have the necessary sequence_number field
2025-08-13 16:31:25 -07:00
..
__init__.py chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
agent_instance.py chore: standardize session not found error (#3031) 2025-08-04 13:12:02 -07:00
agents.py feat(responses): add include parameter (#3115) 2025-08-12 10:24:01 -07:00
config.py feat: add list responses API (#2233) 2025-05-23 13:16:48 -07:00
openai_responses.py feat(responses): stream progress of tool calls (#3135) 2025-08-13 16:31:25 -07:00
persistence.py fix: Fix list_sessions() (#3114) 2025-08-13 07:46:26 -07:00
safety.py chore(api): add mypy coverage to meta_reference_safety (#2661) 2025-07-09 10:22:34 +02:00