diff --git a/docs/_static/llama-stack-spec.html b/docs/_static/llama-stack-spec.html
index 8d0e8652e..ae599563d 100644
--- a/docs/_static/llama-stack-spec.html
+++ b/docs/_static/llama-stack-spec.html
@@ -2440,8 +2440,7 @@
"required": [
"role",
"content",
- "stop_reason",
- "tool_calls"
+ "stop_reason"
],
"description": "A message containing the model's (assistant) response in a chat conversation."
},
@@ -3632,7 +3631,6 @@
},
"additionalProperties": false,
"required": [
- "max_infer_iters",
"model",
"instructions",
"enable_session_persistence"
@@ -4217,7 +4215,6 @@
"input_messages",
"steps",
"output_message",
- "output_attachments",
"started_at"
],
"description": "A single turn in an interaction with an Agentic System."
diff --git a/docs/_static/llama-stack-spec.yaml b/docs/_static/llama-stack-spec.yaml
index dbae7dbb2..2953f1b69 100644
--- a/docs/_static/llama-stack-spec.yaml
+++ b/docs/_static/llama-stack-spec.yaml
@@ -1486,7 +1486,6 @@ components:
- role
- content
- stop_reason
- - tool_calls
description: >-
A message containing the model's (assistant) response in a chat conversation.
GrammarResponseFormat:
@@ -2361,7 +2360,6 @@ components:
$ref: '#/components/schemas/ResponseFormat'
additionalProperties: false
required:
- - max_infer_iters
- model
- instructions
- enable_session_persistence
@@ -2718,7 +2716,6 @@ components:
- input_messages
- steps
- output_message
- - output_attachments
- started_at
description: >-
A single turn in an interaction with an Agentic System.
diff --git a/llama_stack/apis/agents/agents.py b/llama_stack/apis/agents/agents.py
index 50bea3d55..38c6b5561 100644
--- a/llama_stack/apis/agents/agents.py
+++ b/llama_stack/apis/agents/agents.py
@@ -116,7 +116,7 @@ class Turn(BaseModel):
]
steps: List[Step]
output_message: CompletionMessage
- output_attachments: List[Attachment] = Field(default_factory=list)
+ output_attachments: Optional[List[Attachment]] = Field(default_factory=list)
started_at: datetime
completed_at: Optional[datetime] = None
@@ -156,7 +156,7 @@ class AgentConfigCommon(BaseModel):
tool_choice: Optional[ToolChoice] = Field(default=ToolChoice.auto)
tool_prompt_format: Optional[ToolPromptFormat] = Field(default=None)
- max_infer_iters: int = 10
+ max_infer_iters: Optional[int] = 10
@json_schema_type
diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py
index d4f13d65c..4e095e831 100644
--- a/llama_stack/apis/inference/inference.py
+++ b/llama_stack/apis/inference/inference.py
@@ -142,7 +142,7 @@ class CompletionMessage(BaseModel):
role: Literal["assistant"] = "assistant"
content: InterleavedContent
stop_reason: StopReason
- tool_calls: List[ToolCall] = Field(default_factory=list)
+ tool_calls: Optional[List[ToolCall]] = Field(default_factory=list)
Message = register_schema(