mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 12:06:04 +00:00
feat(responses)!: add support for OpenAI compatible Prompts in Responses API
This commit is contained in:
parent
bd3c473208
commit
59169bfd25
33 changed files with 1667 additions and 34 deletions
81
docs/static/stainless-llama-stack-spec.html
vendored
81
docs/static/stainless-llama-stack-spec.html
vendored
|
|
@ -7401,16 +7401,53 @@
|
|||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
],
|
||||
"discriminator": {
|
||||
"propertyName": "type",
|
||||
"mapping": {
|
||||
"input_text": "#/components/schemas/OpenAIResponseInputMessageContentText",
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage",
|
||||
"input_file": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
}
|
||||
},
|
||||
"OpenAIResponseInputMessageContentFile": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"type": {
|
||||
"type": "string",
|
||||
"const": "input_file",
|
||||
"default": "input_file",
|
||||
"description": "The type of the input item. Always `input_file`."
|
||||
},
|
||||
"file_data": {
|
||||
"type": "string",
|
||||
"description": "The data of the file to be sent to the model."
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"file_url": {
|
||||
"type": "string",
|
||||
"description": "The URL of the file to be sent to the model."
|
||||
},
|
||||
"filename": {
|
||||
"type": "string",
|
||||
"description": "The name of the file to be sent to the model."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"type"
|
||||
],
|
||||
"title": "OpenAIResponseInputMessageContentFile",
|
||||
"description": "File content for input messages in OpenAI response format."
|
||||
},
|
||||
"OpenAIResponseInputMessageContentImage": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -7438,6 +7475,10 @@
|
|||
"default": "input_image",
|
||||
"description": "Content type identifier, always \"input_image\""
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"image_url": {
|
||||
"type": "string",
|
||||
"description": "(Optional) URL of the image content"
|
||||
|
|
@ -9241,6 +9282,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
@ -9685,6 +9730,32 @@
|
|||
"title": "OpenAIResponseInputToolMCP",
|
||||
"description": "Model Context Protocol (MCP) tool configuration for OpenAI response inputs."
|
||||
},
|
||||
"OpenAIResponsePromptParam": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"id": {
|
||||
"type": "string",
|
||||
"description": "Unique identifier of the prompt template"
|
||||
},
|
||||
"variables": {
|
||||
"type": "object",
|
||||
"additionalProperties": {
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContent"
|
||||
},
|
||||
"description": "Dictionary of variable names to OpenAIResponseInputMessageContent structure for template substitution"
|
||||
},
|
||||
"version": {
|
||||
"type": "string",
|
||||
"description": "Version number of the prompt to use (defaults to latest if not specified)"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"id"
|
||||
],
|
||||
"title": "OpenAIResponsePromptParam",
|
||||
"description": "Prompt object that is used for OpenAI responses."
|
||||
},
|
||||
"CreateOpenaiResponseRequest": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -9706,6 +9777,10 @@
|
|||
"type": "string",
|
||||
"description": "The underlying LLM used for completions."
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/OpenAIResponsePromptParam",
|
||||
"description": "Prompt object with ID, version, and variables."
|
||||
},
|
||||
"instructions": {
|
||||
"type": "string"
|
||||
},
|
||||
|
|
@ -9794,6 +9869,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue