diff --git a/docs/_static/llama-stack-spec.html b/docs/_static/llama-stack-spec.html
index 54d888441..24fde9054 100644
--- a/docs/_static/llama-stack-spec.html
+++ b/docs/_static/llama-stack-spec.html
@@ -8891,8 +8891,7 @@
},
"additionalProperties": false,
"required": [
- "role",
- "content"
+ "role"
],
"title": "OpenAIAssistantMessageParam",
"description": "A message containing the model's (assistant) response in an OpenAI-compatible chat completion request."
diff --git a/docs/_static/llama-stack-spec.yaml b/docs/_static/llama-stack-spec.yaml
index cf657bff9..27712ee74 100644
--- a/docs/_static/llama-stack-spec.yaml
+++ b/docs/_static/llama-stack-spec.yaml
@@ -6097,7 +6097,6 @@ components:
additionalProperties: false
required:
- role
- - content
title: OpenAIAssistantMessageParam
description: >-
A message containing the model's (assistant) response in an OpenAI-compatible
diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py
index 596efb136..309171f20 100644
--- a/llama_stack/apis/inference/inference.py
+++ b/llama_stack/apis/inference/inference.py
@@ -526,9 +526,9 @@ class OpenAIAssistantMessageParam(BaseModel):
"""
role: Literal["assistant"] = "assistant"
- content: OpenAIChatCompletionMessageContent
+ content: Optional[OpenAIChatCompletionMessageContent] = None
name: Optional[str] = None
- tool_calls: Optional[List[OpenAIChatCompletionToolCall]] = Field(default_factory=list)
+ tool_calls: Optional[List[OpenAIChatCompletionToolCall]] = None
@json_schema_type