diff --git a/docs/static/llama-stack-spec.html b/docs/static/llama-stack-spec.html
index cd64117fd..c66ff458a 100644
--- a/docs/static/llama-stack-spec.html
+++ b/docs/static/llama-stack-spec.html
@@ -496,7 +496,18 @@
"schema": {
"type": "array",
"items": {
- "$ref": "#/components/schemas/ConversationItemInclude"
+ "type": "string",
+ "enum": [
+ "web_search_call.action.sources",
+ "code_interpreter_call.outputs",
+ "computer_call_output.output.image_url",
+ "file_search_call.results",
+ "message.input_image.image_url",
+ "message.output_text.logprobs",
+ "reasoning.encrypted_content"
+ ],
+ "title": "ConversationItemInclude",
+ "description": "Specify additional output data to include in the model response."
}
}
},
@@ -6311,20 +6322,6 @@
"title": "ConversationDeletedResource",
"description": "Response for deleted conversation."
},
- "ConversationItemInclude": {
- "type": "string",
- "enum": [
- "web_search_call.action.sources",
- "code_interpreter_call.outputs",
- "computer_call_output.output.image_url",
- "file_search_call.results",
- "message.input_image.image_url",
- "message.output_text.logprobs",
- "reasoning.encrypted_content"
- ],
- "title": "ConversationItemInclude",
- "description": "Specify additional output data to include in the model response."
- },
"ConversationItemList": {
"type": "object",
"properties": {
diff --git a/docs/static/llama-stack-spec.yaml b/docs/static/llama-stack-spec.yaml
index 670138c4c..5166d05c6 100644
--- a/docs/static/llama-stack-spec.yaml
+++ b/docs/static/llama-stack-spec.yaml
@@ -358,7 +358,18 @@ paths:
schema:
type: array
items:
- $ref: '#/components/schemas/ConversationItemInclude'
+ type: string
+ enum:
+ - web_search_call.action.sources
+ - code_interpreter_call.outputs
+ - computer_call_output.output.image_url
+ - file_search_call.results
+ - message.input_image.image_url
+ - message.output_text.logprobs
+ - reasoning.encrypted_content
+ title: ConversationItemInclude
+ description: >-
+ Specify additional output data to include in the model response.
- name: limit
in: query
description: >-
@@ -4738,19 +4749,6 @@ components:
- deleted
title: ConversationDeletedResource
description: Response for deleted conversation.
- ConversationItemInclude:
- type: string
- enum:
- - web_search_call.action.sources
- - code_interpreter_call.outputs
- - computer_call_output.output.image_url
- - file_search_call.results
- - message.input_image.image_url
- - message.output_text.logprobs
- - reasoning.encrypted_content
- title: ConversationItemInclude
- description: >-
- Specify additional output data to include in the model response.
ConversationItemList:
type: object
properties:
diff --git a/docs/static/stainless-llama-stack-spec.html b/docs/static/stainless-llama-stack-spec.html
index 31433f76d..171f8dc2c 100644
--- a/docs/static/stainless-llama-stack-spec.html
+++ b/docs/static/stainless-llama-stack-spec.html
@@ -496,7 +496,18 @@
"schema": {
"type": "array",
"items": {
- "$ref": "#/components/schemas/ConversationItemInclude"
+ "type": "string",
+ "enum": [
+ "web_search_call.action.sources",
+ "code_interpreter_call.outputs",
+ "computer_call_output.output.image_url",
+ "file_search_call.results",
+ "message.input_image.image_url",
+ "message.output_text.logprobs",
+ "reasoning.encrypted_content"
+ ],
+ "title": "ConversationItemInclude",
+ "description": "Specify additional output data to include in the model response."
}
}
},
@@ -7983,20 +7994,6 @@
"title": "ConversationDeletedResource",
"description": "Response for deleted conversation."
},
- "ConversationItemInclude": {
- "type": "string",
- "enum": [
- "web_search_call.action.sources",
- "code_interpreter_call.outputs",
- "computer_call_output.output.image_url",
- "file_search_call.results",
- "message.input_image.image_url",
- "message.output_text.logprobs",
- "reasoning.encrypted_content"
- ],
- "title": "ConversationItemInclude",
- "description": "Specify additional output data to include in the model response."
- },
"ConversationItemList": {
"type": "object",
"properties": {
diff --git a/docs/static/stainless-llama-stack-spec.yaml b/docs/static/stainless-llama-stack-spec.yaml
index 2b0436946..cb9bac130 100644
--- a/docs/static/stainless-llama-stack-spec.yaml
+++ b/docs/static/stainless-llama-stack-spec.yaml
@@ -361,7 +361,18 @@ paths:
schema:
type: array
items:
- $ref: '#/components/schemas/ConversationItemInclude'
+ type: string
+ enum:
+ - web_search_call.action.sources
+ - code_interpreter_call.outputs
+ - computer_call_output.output.image_url
+ - file_search_call.results
+ - message.input_image.image_url
+ - message.output_text.logprobs
+ - reasoning.encrypted_content
+ title: ConversationItemInclude
+ description: >-
+ Specify additional output data to include in the model response.
- name: limit
in: query
description: >-
@@ -5951,19 +5962,6 @@ components:
- deleted
title: ConversationDeletedResource
description: Response for deleted conversation.
- ConversationItemInclude:
- type: string
- enum:
- - web_search_call.action.sources
- - code_interpreter_call.outputs
- - computer_call_output.output.image_url
- - file_search_call.results
- - message.input_image.image_url
- - message.output_text.logprobs
- - reasoning.encrypted_content
- title: ConversationItemInclude
- description: >-
- Specify additional output data to include in the model response.
ConversationItemList:
type: object
properties:
diff --git a/llama_stack/apis/conversations/conversations.py b/llama_stack/apis/conversations/conversations.py
index 2adbd157a..3b6c50a03 100644
--- a/llama_stack/apis/conversations/conversations.py
+++ b/llama_stack/apis/conversations/conversations.py
@@ -148,7 +148,6 @@ class ConversationItemCreateRequest(BaseModel):
)
-@json_schema_type
class ConversationItemInclude(StrEnum):
"""
Specify additional output data to include in the model response.
diff --git a/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py b/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py
index 851e6ef28..684237512 100644
--- a/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py
+++ b/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py
@@ -131,7 +131,7 @@ class OpenAIResponsesImpl:
tool_context.recover_tools_from_previous_response(previous_response)
elif conversation is not None:
- conversation_items = await self.conversations_api.list(conversation, order="asc")
+ conversation_items = await self.conversations_api.list_items(conversation, order="asc")
# Use stored messages as source of truth (like previous_response.messages)
stored_messages = await self.responses_store.get_conversation_messages(conversation)