mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 12:06:04 +00:00
feat(responses)!: add support for OpenAI compatible Prompts in Responses API
This commit is contained in:
parent
bd3c473208
commit
59169bfd25
33 changed files with 1667 additions and 34 deletions
119
docs/static/deprecated-llama-stack-spec.html
vendored
119
docs/static/deprecated-llama-stack-spec.html
vendored
|
|
@ -8593,16 +8593,53 @@
|
|||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
],
|
||||
"discriminator": {
|
||||
"propertyName": "type",
|
||||
"mapping": {
|
||||
"input_text": "#/components/schemas/OpenAIResponseInputMessageContentText",
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage",
|
||||
"input_file": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
}
|
||||
},
|
||||
"OpenAIResponseInputMessageContentFile": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"type": {
|
||||
"type": "string",
|
||||
"const": "input_file",
|
||||
"default": "input_file",
|
||||
"description": "The type of the input item. Always `input_file`."
|
||||
},
|
||||
"file_data": {
|
||||
"type": "string",
|
||||
"description": "The data of the file to be sent to the model."
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"file_url": {
|
||||
"type": "string",
|
||||
"description": "The URL of the file to be sent to the model."
|
||||
},
|
||||
"filename": {
|
||||
"type": "string",
|
||||
"description": "The name of the file to be sent to the model."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"type"
|
||||
],
|
||||
"title": "OpenAIResponseInputMessageContentFile",
|
||||
"description": "File content for input messages in OpenAI response format."
|
||||
},
|
||||
"OpenAIResponseInputMessageContentImage": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -8630,6 +8667,10 @@
|
|||
"default": "input_image",
|
||||
"description": "Content type identifier, always \"input_image\""
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"image_url": {
|
||||
"type": "string",
|
||||
"description": "(Optional) URL of the image content"
|
||||
|
|
@ -8993,6 +9034,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
@ -9610,6 +9655,44 @@
|
|||
"title": "OpenAIResponseUsage",
|
||||
"description": "Usage information for OpenAI response."
|
||||
},
|
||||
"Prompt": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"prompt": {
|
||||
"type": "string",
|
||||
"description": "The system prompt text with variable placeholders. Variables are only supported when using the Responses API."
|
||||
},
|
||||
"version": {
|
||||
"type": "integer",
|
||||
"description": "Version (integer starting at 1, incremented on save)"
|
||||
},
|
||||
"prompt_id": {
|
||||
"type": "string",
|
||||
"description": "Unique identifier formatted as 'pmpt_<48-digit-hash>'"
|
||||
},
|
||||
"variables": {
|
||||
"type": "array",
|
||||
"items": {
|
||||
"type": "string"
|
||||
},
|
||||
"description": "List of prompt variable names that can be used in the prompt template"
|
||||
},
|
||||
"is_default": {
|
||||
"type": "boolean",
|
||||
"default": false,
|
||||
"description": "Boolean indicating whether this version is the default version for this prompt"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"version",
|
||||
"prompt_id",
|
||||
"variables",
|
||||
"is_default"
|
||||
],
|
||||
"title": "Prompt",
|
||||
"description": "A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."
|
||||
},
|
||||
"ResponseGuardrailSpec": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -9766,6 +9849,32 @@
|
|||
"title": "OpenAIResponseInputToolMCP",
|
||||
"description": "Model Context Protocol (MCP) tool configuration for OpenAI response inputs."
|
||||
},
|
||||
"OpenAIResponsePromptParam": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"id": {
|
||||
"type": "string",
|
||||
"description": "Unique identifier of the prompt template"
|
||||
},
|
||||
"variables": {
|
||||
"type": "object",
|
||||
"additionalProperties": {
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContent"
|
||||
},
|
||||
"description": "Dictionary of variable names to OpenAIResponseInputMessageContent structure for template substitution"
|
||||
},
|
||||
"version": {
|
||||
"type": "string",
|
||||
"description": "Version number of the prompt to use (defaults to latest if not specified)"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"id"
|
||||
],
|
||||
"title": "OpenAIResponsePromptParam",
|
||||
"description": "Prompt object that is used for OpenAI responses."
|
||||
},
|
||||
"CreateOpenaiResponseRequest": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -9787,6 +9896,10 @@
|
|||
"type": "string",
|
||||
"description": "The underlying LLM used for completions."
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/OpenAIResponsePromptParam",
|
||||
"description": "Prompt object with ID, version, and variables."
|
||||
},
|
||||
"instructions": {
|
||||
"type": "string"
|
||||
},
|
||||
|
|
@ -9875,6 +9988,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
|
|||
110
docs/static/deprecated-llama-stack-spec.yaml
vendored
110
docs/static/deprecated-llama-stack-spec.yaml
vendored
|
|
@ -6409,11 +6409,44 @@ components:
|
|||
oneOf:
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentText'
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentImage'
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentFile'
|
||||
discriminator:
|
||||
propertyName: type
|
||||
mapping:
|
||||
input_text: '#/components/schemas/OpenAIResponseInputMessageContentText'
|
||||
input_image: '#/components/schemas/OpenAIResponseInputMessageContentImage'
|
||||
input_file: '#/components/schemas/OpenAIResponseInputMessageContentFile'
|
||||
OpenAIResponseInputMessageContentFile:
|
||||
type: object
|
||||
properties:
|
||||
type:
|
||||
type: string
|
||||
const: input_file
|
||||
default: input_file
|
||||
description: >-
|
||||
The type of the input item. Always `input_file`.
|
||||
file_data:
|
||||
type: string
|
||||
description: >-
|
||||
The data of the file to be sent to the model.
|
||||
file_id:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) The ID of the file to be sent to the model.
|
||||
file_url:
|
||||
type: string
|
||||
description: >-
|
||||
The URL of the file to be sent to the model.
|
||||
filename:
|
||||
type: string
|
||||
description: >-
|
||||
The name of the file to be sent to the model.
|
||||
additionalProperties: false
|
||||
required:
|
||||
- type
|
||||
title: OpenAIResponseInputMessageContentFile
|
||||
description: >-
|
||||
File content for input messages in OpenAI response format.
|
||||
OpenAIResponseInputMessageContentImage:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -6434,6 +6467,10 @@ components:
|
|||
default: input_image
|
||||
description: >-
|
||||
Content type identifier, always "input_image"
|
||||
file_id:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) The ID of the file to be sent to the model.
|
||||
image_url:
|
||||
type: string
|
||||
description: (Optional) URL of the image content
|
||||
|
|
@ -6704,6 +6741,10 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) ID of the previous response in a conversation
|
||||
prompt:
|
||||
$ref: '#/components/schemas/Prompt'
|
||||
description: >-
|
||||
(Optional) Prompt object with ID, version, and variables
|
||||
status:
|
||||
type: string
|
||||
description: >-
|
||||
|
|
@ -7181,6 +7222,44 @@ components:
|
|||
- total_tokens
|
||||
title: OpenAIResponseUsage
|
||||
description: Usage information for OpenAI response.
|
||||
Prompt:
|
||||
type: object
|
||||
properties:
|
||||
prompt:
|
||||
type: string
|
||||
description: >-
|
||||
The system prompt text with variable placeholders. Variables are only
|
||||
supported when using the Responses API.
|
||||
version:
|
||||
type: integer
|
||||
description: >-
|
||||
Version (integer starting at 1, incremented on save)
|
||||
prompt_id:
|
||||
type: string
|
||||
description: >-
|
||||
Unique identifier formatted as 'pmpt_<48-digit-hash>'
|
||||
variables:
|
||||
type: array
|
||||
items:
|
||||
type: string
|
||||
description: >-
|
||||
List of prompt variable names that can be used in the prompt template
|
||||
is_default:
|
||||
type: boolean
|
||||
default: false
|
||||
description: >-
|
||||
Boolean indicating whether this version is the default version for this
|
||||
prompt
|
||||
additionalProperties: false
|
||||
required:
|
||||
- version
|
||||
- prompt_id
|
||||
- variables
|
||||
- is_default
|
||||
title: Prompt
|
||||
description: >-
|
||||
A prompt resource representing a stored OpenAI Compatible prompt template
|
||||
in Llama Stack.
|
||||
ResponseGuardrailSpec:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -7287,6 +7366,29 @@ components:
|
|||
title: OpenAIResponseInputToolMCP
|
||||
description: >-
|
||||
Model Context Protocol (MCP) tool configuration for OpenAI response inputs.
|
||||
OpenAIResponsePromptParam:
|
||||
type: object
|
||||
properties:
|
||||
id:
|
||||
type: string
|
||||
description: Unique identifier of the prompt template
|
||||
variables:
|
||||
type: object
|
||||
additionalProperties:
|
||||
$ref: '#/components/schemas/OpenAIResponseInputMessageContent'
|
||||
description: >-
|
||||
Dictionary of variable names to OpenAIResponseInputMessageContent structure
|
||||
for template substitution
|
||||
version:
|
||||
type: string
|
||||
description: >-
|
||||
Version number of the prompt to use (defaults to latest if not specified)
|
||||
additionalProperties: false
|
||||
required:
|
||||
- id
|
||||
title: OpenAIResponsePromptParam
|
||||
description: >-
|
||||
Prompt object that is used for OpenAI responses.
|
||||
CreateOpenaiResponseRequest:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -7300,6 +7402,10 @@ components:
|
|||
model:
|
||||
type: string
|
||||
description: The underlying LLM used for completions.
|
||||
prompt:
|
||||
$ref: '#/components/schemas/OpenAIResponsePromptParam'
|
||||
description: >-
|
||||
Prompt object with ID, version, and variables.
|
||||
instructions:
|
||||
type: string
|
||||
previous_response_id:
|
||||
|
|
@ -7377,6 +7483,10 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) ID of the previous response in a conversation
|
||||
prompt:
|
||||
$ref: '#/components/schemas/Prompt'
|
||||
description: >-
|
||||
(Optional) Prompt object with ID, version, and variables
|
||||
status:
|
||||
type: string
|
||||
description: >-
|
||||
|
|
|
|||
81
docs/static/llama-stack-spec.html
vendored
81
docs/static/llama-stack-spec.html
vendored
|
|
@ -5729,16 +5729,53 @@
|
|||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
],
|
||||
"discriminator": {
|
||||
"propertyName": "type",
|
||||
"mapping": {
|
||||
"input_text": "#/components/schemas/OpenAIResponseInputMessageContentText",
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage",
|
||||
"input_file": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
}
|
||||
},
|
||||
"OpenAIResponseInputMessageContentFile": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"type": {
|
||||
"type": "string",
|
||||
"const": "input_file",
|
||||
"default": "input_file",
|
||||
"description": "The type of the input item. Always `input_file`."
|
||||
},
|
||||
"file_data": {
|
||||
"type": "string",
|
||||
"description": "The data of the file to be sent to the model."
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"file_url": {
|
||||
"type": "string",
|
||||
"description": "The URL of the file to be sent to the model."
|
||||
},
|
||||
"filename": {
|
||||
"type": "string",
|
||||
"description": "The name of the file to be sent to the model."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"type"
|
||||
],
|
||||
"title": "OpenAIResponseInputMessageContentFile",
|
||||
"description": "File content for input messages in OpenAI response format."
|
||||
},
|
||||
"OpenAIResponseInputMessageContentImage": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -5766,6 +5803,10 @@
|
|||
"default": "input_image",
|
||||
"description": "Content type identifier, always \"input_image\""
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"image_url": {
|
||||
"type": "string",
|
||||
"description": "(Optional) URL of the image content"
|
||||
|
|
@ -7569,6 +7610,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
@ -8013,6 +8058,32 @@
|
|||
"title": "OpenAIResponseInputToolMCP",
|
||||
"description": "Model Context Protocol (MCP) tool configuration for OpenAI response inputs."
|
||||
},
|
||||
"OpenAIResponsePromptParam": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"id": {
|
||||
"type": "string",
|
||||
"description": "Unique identifier of the prompt template"
|
||||
},
|
||||
"variables": {
|
||||
"type": "object",
|
||||
"additionalProperties": {
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContent"
|
||||
},
|
||||
"description": "Dictionary of variable names to OpenAIResponseInputMessageContent structure for template substitution"
|
||||
},
|
||||
"version": {
|
||||
"type": "string",
|
||||
"description": "Version number of the prompt to use (defaults to latest if not specified)"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"id"
|
||||
],
|
||||
"title": "OpenAIResponsePromptParam",
|
||||
"description": "Prompt object that is used for OpenAI responses."
|
||||
},
|
||||
"CreateOpenaiResponseRequest": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -8034,6 +8105,10 @@
|
|||
"type": "string",
|
||||
"description": "The underlying LLM used for completions."
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/OpenAIResponsePromptParam",
|
||||
"description": "Prompt object with ID, version, and variables."
|
||||
},
|
||||
"instructions": {
|
||||
"type": "string"
|
||||
},
|
||||
|
|
@ -8122,6 +8197,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
|
|||
72
docs/static/llama-stack-spec.yaml
vendored
72
docs/static/llama-stack-spec.yaml
vendored
|
|
@ -4361,11 +4361,44 @@ components:
|
|||
oneOf:
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentText'
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentImage'
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentFile'
|
||||
discriminator:
|
||||
propertyName: type
|
||||
mapping:
|
||||
input_text: '#/components/schemas/OpenAIResponseInputMessageContentText'
|
||||
input_image: '#/components/schemas/OpenAIResponseInputMessageContentImage'
|
||||
input_file: '#/components/schemas/OpenAIResponseInputMessageContentFile'
|
||||
OpenAIResponseInputMessageContentFile:
|
||||
type: object
|
||||
properties:
|
||||
type:
|
||||
type: string
|
||||
const: input_file
|
||||
default: input_file
|
||||
description: >-
|
||||
The type of the input item. Always `input_file`.
|
||||
file_data:
|
||||
type: string
|
||||
description: >-
|
||||
The data of the file to be sent to the model.
|
||||
file_id:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) The ID of the file to be sent to the model.
|
||||
file_url:
|
||||
type: string
|
||||
description: >-
|
||||
The URL of the file to be sent to the model.
|
||||
filename:
|
||||
type: string
|
||||
description: >-
|
||||
The name of the file to be sent to the model.
|
||||
additionalProperties: false
|
||||
required:
|
||||
- type
|
||||
title: OpenAIResponseInputMessageContentFile
|
||||
description: >-
|
||||
File content for input messages in OpenAI response format.
|
||||
OpenAIResponseInputMessageContentImage:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -4386,6 +4419,10 @@ components:
|
|||
default: input_image
|
||||
description: >-
|
||||
Content type identifier, always "input_image"
|
||||
file_id:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) The ID of the file to be sent to the model.
|
||||
image_url:
|
||||
type: string
|
||||
description: (Optional) URL of the image content
|
||||
|
|
@ -5785,6 +5822,10 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) ID of the previous response in a conversation
|
||||
prompt:
|
||||
$ref: '#/components/schemas/Prompt'
|
||||
description: >-
|
||||
(Optional) Prompt object with ID, version, and variables
|
||||
status:
|
||||
type: string
|
||||
description: >-
|
||||
|
|
@ -6102,6 +6143,29 @@ components:
|
|||
title: OpenAIResponseInputToolMCP
|
||||
description: >-
|
||||
Model Context Protocol (MCP) tool configuration for OpenAI response inputs.
|
||||
OpenAIResponsePromptParam:
|
||||
type: object
|
||||
properties:
|
||||
id:
|
||||
type: string
|
||||
description: Unique identifier of the prompt template
|
||||
variables:
|
||||
type: object
|
||||
additionalProperties:
|
||||
$ref: '#/components/schemas/OpenAIResponseInputMessageContent'
|
||||
description: >-
|
||||
Dictionary of variable names to OpenAIResponseInputMessageContent structure
|
||||
for template substitution
|
||||
version:
|
||||
type: string
|
||||
description: >-
|
||||
Version number of the prompt to use (defaults to latest if not specified)
|
||||
additionalProperties: false
|
||||
required:
|
||||
- id
|
||||
title: OpenAIResponsePromptParam
|
||||
description: >-
|
||||
Prompt object that is used for OpenAI responses.
|
||||
CreateOpenaiResponseRequest:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -6115,6 +6179,10 @@ components:
|
|||
model:
|
||||
type: string
|
||||
description: The underlying LLM used for completions.
|
||||
prompt:
|
||||
$ref: '#/components/schemas/OpenAIResponsePromptParam'
|
||||
description: >-
|
||||
Prompt object with ID, version, and variables.
|
||||
instructions:
|
||||
type: string
|
||||
previous_response_id:
|
||||
|
|
@ -6192,6 +6260,10 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) ID of the previous response in a conversation
|
||||
prompt:
|
||||
$ref: '#/components/schemas/Prompt'
|
||||
description: >-
|
||||
(Optional) Prompt object with ID, version, and variables
|
||||
status:
|
||||
type: string
|
||||
description: >-
|
||||
|
|
|
|||
81
docs/static/stainless-llama-stack-spec.html
vendored
81
docs/static/stainless-llama-stack-spec.html
vendored
|
|
@ -7401,16 +7401,53 @@
|
|||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
},
|
||||
{
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
],
|
||||
"discriminator": {
|
||||
"propertyName": "type",
|
||||
"mapping": {
|
||||
"input_text": "#/components/schemas/OpenAIResponseInputMessageContentText",
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage"
|
||||
"input_image": "#/components/schemas/OpenAIResponseInputMessageContentImage",
|
||||
"input_file": "#/components/schemas/OpenAIResponseInputMessageContentFile"
|
||||
}
|
||||
}
|
||||
},
|
||||
"OpenAIResponseInputMessageContentFile": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"type": {
|
||||
"type": "string",
|
||||
"const": "input_file",
|
||||
"default": "input_file",
|
||||
"description": "The type of the input item. Always `input_file`."
|
||||
},
|
||||
"file_data": {
|
||||
"type": "string",
|
||||
"description": "The data of the file to be sent to the model."
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"file_url": {
|
||||
"type": "string",
|
||||
"description": "The URL of the file to be sent to the model."
|
||||
},
|
||||
"filename": {
|
||||
"type": "string",
|
||||
"description": "The name of the file to be sent to the model."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"type"
|
||||
],
|
||||
"title": "OpenAIResponseInputMessageContentFile",
|
||||
"description": "File content for input messages in OpenAI response format."
|
||||
},
|
||||
"OpenAIResponseInputMessageContentImage": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -7438,6 +7475,10 @@
|
|||
"default": "input_image",
|
||||
"description": "Content type identifier, always \"input_image\""
|
||||
},
|
||||
"file_id": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The ID of the file to be sent to the model."
|
||||
},
|
||||
"image_url": {
|
||||
"type": "string",
|
||||
"description": "(Optional) URL of the image content"
|
||||
|
|
@ -9241,6 +9282,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
@ -9685,6 +9730,32 @@
|
|||
"title": "OpenAIResponseInputToolMCP",
|
||||
"description": "Model Context Protocol (MCP) tool configuration for OpenAI response inputs."
|
||||
},
|
||||
"OpenAIResponsePromptParam": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"id": {
|
||||
"type": "string",
|
||||
"description": "Unique identifier of the prompt template"
|
||||
},
|
||||
"variables": {
|
||||
"type": "object",
|
||||
"additionalProperties": {
|
||||
"$ref": "#/components/schemas/OpenAIResponseInputMessageContent"
|
||||
},
|
||||
"description": "Dictionary of variable names to OpenAIResponseInputMessageContent structure for template substitution"
|
||||
},
|
||||
"version": {
|
||||
"type": "string",
|
||||
"description": "Version number of the prompt to use (defaults to latest if not specified)"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"id"
|
||||
],
|
||||
"title": "OpenAIResponsePromptParam",
|
||||
"description": "Prompt object that is used for OpenAI responses."
|
||||
},
|
||||
"CreateOpenaiResponseRequest": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
|
@ -9706,6 +9777,10 @@
|
|||
"type": "string",
|
||||
"description": "The underlying LLM used for completions."
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/OpenAIResponsePromptParam",
|
||||
"description": "Prompt object with ID, version, and variables."
|
||||
},
|
||||
"instructions": {
|
||||
"type": "string"
|
||||
},
|
||||
|
|
@ -9794,6 +9869,10 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) ID of the previous response in a conversation"
|
||||
},
|
||||
"prompt": {
|
||||
"$ref": "#/components/schemas/Prompt",
|
||||
"description": "(Optional) Prompt object with ID, version, and variables"
|
||||
},
|
||||
"status": {
|
||||
"type": "string",
|
||||
"description": "Current status of the response generation"
|
||||
|
|
|
|||
72
docs/static/stainless-llama-stack-spec.yaml
vendored
72
docs/static/stainless-llama-stack-spec.yaml
vendored
|
|
@ -5574,11 +5574,44 @@ components:
|
|||
oneOf:
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentText'
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentImage'
|
||||
- $ref: '#/components/schemas/OpenAIResponseInputMessageContentFile'
|
||||
discriminator:
|
||||
propertyName: type
|
||||
mapping:
|
||||
input_text: '#/components/schemas/OpenAIResponseInputMessageContentText'
|
||||
input_image: '#/components/schemas/OpenAIResponseInputMessageContentImage'
|
||||
input_file: '#/components/schemas/OpenAIResponseInputMessageContentFile'
|
||||
OpenAIResponseInputMessageContentFile:
|
||||
type: object
|
||||
properties:
|
||||
type:
|
||||
type: string
|
||||
const: input_file
|
||||
default: input_file
|
||||
description: >-
|
||||
The type of the input item. Always `input_file`.
|
||||
file_data:
|
||||
type: string
|
||||
description: >-
|
||||
The data of the file to be sent to the model.
|
||||
file_id:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) The ID of the file to be sent to the model.
|
||||
file_url:
|
||||
type: string
|
||||
description: >-
|
||||
The URL of the file to be sent to the model.
|
||||
filename:
|
||||
type: string
|
||||
description: >-
|
||||
The name of the file to be sent to the model.
|
||||
additionalProperties: false
|
||||
required:
|
||||
- type
|
||||
title: OpenAIResponseInputMessageContentFile
|
||||
description: >-
|
||||
File content for input messages in OpenAI response format.
|
||||
OpenAIResponseInputMessageContentImage:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -5599,6 +5632,10 @@ components:
|
|||
default: input_image
|
||||
description: >-
|
||||
Content type identifier, always "input_image"
|
||||
file_id:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) The ID of the file to be sent to the model.
|
||||
image_url:
|
||||
type: string
|
||||
description: (Optional) URL of the image content
|
||||
|
|
@ -6998,6 +7035,10 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) ID of the previous response in a conversation
|
||||
prompt:
|
||||
$ref: '#/components/schemas/Prompt'
|
||||
description: >-
|
||||
(Optional) Prompt object with ID, version, and variables
|
||||
status:
|
||||
type: string
|
||||
description: >-
|
||||
|
|
@ -7315,6 +7356,29 @@ components:
|
|||
title: OpenAIResponseInputToolMCP
|
||||
description: >-
|
||||
Model Context Protocol (MCP) tool configuration for OpenAI response inputs.
|
||||
OpenAIResponsePromptParam:
|
||||
type: object
|
||||
properties:
|
||||
id:
|
||||
type: string
|
||||
description: Unique identifier of the prompt template
|
||||
variables:
|
||||
type: object
|
||||
additionalProperties:
|
||||
$ref: '#/components/schemas/OpenAIResponseInputMessageContent'
|
||||
description: >-
|
||||
Dictionary of variable names to OpenAIResponseInputMessageContent structure
|
||||
for template substitution
|
||||
version:
|
||||
type: string
|
||||
description: >-
|
||||
Version number of the prompt to use (defaults to latest if not specified)
|
||||
additionalProperties: false
|
||||
required:
|
||||
- id
|
||||
title: OpenAIResponsePromptParam
|
||||
description: >-
|
||||
Prompt object that is used for OpenAI responses.
|
||||
CreateOpenaiResponseRequest:
|
||||
type: object
|
||||
properties:
|
||||
|
|
@ -7328,6 +7392,10 @@ components:
|
|||
model:
|
||||
type: string
|
||||
description: The underlying LLM used for completions.
|
||||
prompt:
|
||||
$ref: '#/components/schemas/OpenAIResponsePromptParam'
|
||||
description: >-
|
||||
Prompt object with ID, version, and variables.
|
||||
instructions:
|
||||
type: string
|
||||
previous_response_id:
|
||||
|
|
@ -7405,6 +7473,10 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) ID of the previous response in a conversation
|
||||
prompt:
|
||||
$ref: '#/components/schemas/Prompt'
|
||||
description: >-
|
||||
(Optional) Prompt object with ID, version, and variables
|
||||
status:
|
||||
type: string
|
||||
description: >-
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue