From 55ca8a5c509b675a5b6eac372738fda56db304c3 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Wed, 22 Oct 2025 12:00:12 -0700 Subject: [PATCH] fixes --- docs/static/llama-stack-spec.html | 27 +++++++++---------- docs/static/llama-stack-spec.yaml | 26 +++++++++--------- docs/static/stainless-llama-stack-spec.html | 27 +++++++++---------- docs/static/stainless-llama-stack-spec.yaml | 26 +++++++++--------- .../apis/conversations/conversations.py | 1 - .../responses/openai_responses.py | 2 +- 6 files changed, 49 insertions(+), 60 deletions(-) diff --git a/docs/static/llama-stack-spec.html b/docs/static/llama-stack-spec.html index cd64117fd..c66ff458a 100644 --- a/docs/static/llama-stack-spec.html +++ b/docs/static/llama-stack-spec.html @@ -496,7 +496,18 @@ "schema": { "type": "array", "items": { - "$ref": "#/components/schemas/ConversationItemInclude" + "type": "string", + "enum": [ + "web_search_call.action.sources", + "code_interpreter_call.outputs", + "computer_call_output.output.image_url", + "file_search_call.results", + "message.input_image.image_url", + "message.output_text.logprobs", + "reasoning.encrypted_content" + ], + "title": "ConversationItemInclude", + "description": "Specify additional output data to include in the model response." } } }, @@ -6311,20 +6322,6 @@ "title": "ConversationDeletedResource", "description": "Response for deleted conversation." }, - "ConversationItemInclude": { - "type": "string", - "enum": [ - "web_search_call.action.sources", - "code_interpreter_call.outputs", - "computer_call_output.output.image_url", - "file_search_call.results", - "message.input_image.image_url", - "message.output_text.logprobs", - "reasoning.encrypted_content" - ], - "title": "ConversationItemInclude", - "description": "Specify additional output data to include in the model response." - }, "ConversationItemList": { "type": "object", "properties": { diff --git a/docs/static/llama-stack-spec.yaml b/docs/static/llama-stack-spec.yaml index 670138c4c..5166d05c6 100644 --- a/docs/static/llama-stack-spec.yaml +++ b/docs/static/llama-stack-spec.yaml @@ -358,7 +358,18 @@ paths: schema: type: array items: - $ref: '#/components/schemas/ConversationItemInclude' + type: string + enum: + - web_search_call.action.sources + - code_interpreter_call.outputs + - computer_call_output.output.image_url + - file_search_call.results + - message.input_image.image_url + - message.output_text.logprobs + - reasoning.encrypted_content + title: ConversationItemInclude + description: >- + Specify additional output data to include in the model response. - name: limit in: query description: >- @@ -4738,19 +4749,6 @@ components: - deleted title: ConversationDeletedResource description: Response for deleted conversation. - ConversationItemInclude: - type: string - enum: - - web_search_call.action.sources - - code_interpreter_call.outputs - - computer_call_output.output.image_url - - file_search_call.results - - message.input_image.image_url - - message.output_text.logprobs - - reasoning.encrypted_content - title: ConversationItemInclude - description: >- - Specify additional output data to include in the model response. ConversationItemList: type: object properties: diff --git a/docs/static/stainless-llama-stack-spec.html b/docs/static/stainless-llama-stack-spec.html index 31433f76d..171f8dc2c 100644 --- a/docs/static/stainless-llama-stack-spec.html +++ b/docs/static/stainless-llama-stack-spec.html @@ -496,7 +496,18 @@ "schema": { "type": "array", "items": { - "$ref": "#/components/schemas/ConversationItemInclude" + "type": "string", + "enum": [ + "web_search_call.action.sources", + "code_interpreter_call.outputs", + "computer_call_output.output.image_url", + "file_search_call.results", + "message.input_image.image_url", + "message.output_text.logprobs", + "reasoning.encrypted_content" + ], + "title": "ConversationItemInclude", + "description": "Specify additional output data to include in the model response." } } }, @@ -7983,20 +7994,6 @@ "title": "ConversationDeletedResource", "description": "Response for deleted conversation." }, - "ConversationItemInclude": { - "type": "string", - "enum": [ - "web_search_call.action.sources", - "code_interpreter_call.outputs", - "computer_call_output.output.image_url", - "file_search_call.results", - "message.input_image.image_url", - "message.output_text.logprobs", - "reasoning.encrypted_content" - ], - "title": "ConversationItemInclude", - "description": "Specify additional output data to include in the model response." - }, "ConversationItemList": { "type": "object", "properties": { diff --git a/docs/static/stainless-llama-stack-spec.yaml b/docs/static/stainless-llama-stack-spec.yaml index 2b0436946..cb9bac130 100644 --- a/docs/static/stainless-llama-stack-spec.yaml +++ b/docs/static/stainless-llama-stack-spec.yaml @@ -361,7 +361,18 @@ paths: schema: type: array items: - $ref: '#/components/schemas/ConversationItemInclude' + type: string + enum: + - web_search_call.action.sources + - code_interpreter_call.outputs + - computer_call_output.output.image_url + - file_search_call.results + - message.input_image.image_url + - message.output_text.logprobs + - reasoning.encrypted_content + title: ConversationItemInclude + description: >- + Specify additional output data to include in the model response. - name: limit in: query description: >- @@ -5951,19 +5962,6 @@ components: - deleted title: ConversationDeletedResource description: Response for deleted conversation. - ConversationItemInclude: - type: string - enum: - - web_search_call.action.sources - - code_interpreter_call.outputs - - computer_call_output.output.image_url - - file_search_call.results - - message.input_image.image_url - - message.output_text.logprobs - - reasoning.encrypted_content - title: ConversationItemInclude - description: >- - Specify additional output data to include in the model response. ConversationItemList: type: object properties: diff --git a/llama_stack/apis/conversations/conversations.py b/llama_stack/apis/conversations/conversations.py index 2adbd157a..3b6c50a03 100644 --- a/llama_stack/apis/conversations/conversations.py +++ b/llama_stack/apis/conversations/conversations.py @@ -148,7 +148,6 @@ class ConversationItemCreateRequest(BaseModel): ) -@json_schema_type class ConversationItemInclude(StrEnum): """ Specify additional output data to include in the model response. diff --git a/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py b/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py index 851e6ef28..684237512 100644 --- a/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py +++ b/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py @@ -131,7 +131,7 @@ class OpenAIResponsesImpl: tool_context.recover_tools_from_previous_response(previous_response) elif conversation is not None: - conversation_items = await self.conversations_api.list(conversation, order="asc") + conversation_items = await self.conversations_api.list_items(conversation, order="asc") # Use stored messages as source of truth (like previous_response.messages) stored_messages = await self.responses_store.get_conversation_messages(conversation)