llama-stack-mirror/llama_stack/apis/inference
Ashwin Bharambe 3f1f7c3f7f feat(responses): add reasoning and annotation added events
Implements missing streaming events from OpenAI Responses API spec: reasoning text/summary events for o1/o3 models, refusal events for safety moderation, annotation events for citations, and file search streaming events. Added optional reasoning_content field to chat completion chunks to support non-standard provider extensions. Refactored streaming orchestrator to handle new content types via helper methods.
2025-10-11 14:27:19 -07:00
..
__init__.py chore: remove nested imports (#2515) 2025-06-26 08:01:05 +05:30
event_logger.py pre-commit lint 2024-09-28 16:04:41 -07:00
inference.py feat(responses): add reasoning and annotation added events 2025-10-11 14:27:19 -07:00