mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
1.5 KiB
1.5 KiB
Responses API
The Responses API provides OpenAI-compatible functionality with enhanced capabilities for dynamic, stateful interactions.
✅ STABLE: This API is production-ready with backward compatibility guarantees. Recommended for production applications.
✅ Supported Tools
The Responses API supports the following tool types:
web_search
: Search the web for current information and real-time datafile_search
: Search through uploaded files and vector stores- Supports dynamic
vector_store_ids
per call - Compatible with OpenAI file search patterns
- Supports dynamic
function
: Call custom functions with JSON schema validationmcp_tool
: Model Context Protocol integration
✅ Supported Fields & Features
Core Capabilities:
- Dynamic Configuration: Switch models, vector stores, and tools per request without pre-configuration
- Conversation Branching: Use
previous_response_id
to branch conversations and explore different paths - Rich Annotations: Automatic file citations, URL citations, and container file citations
- Status Tracking: Monitor tool call execution status and handle failures gracefully
🚧 Work in Progress
- Full real-time response streaming support
tool_choice
parametermax_tool_calls
parameter- Built-in tools (code interpreter, containers API)
- Safety & guardrails
reasoning
capabilitiesservice_tier
logprobs
max_output_tokens
metadata
handlinginstructions
incomplete_details
background