mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-06 04:34:57 +00:00
fix(openai_responses): OpenAIResponsesObject is not complete
This commit is contained in:
parent
845e405cd0
commit
17a06452e7
1 changed files with 14 additions and 14 deletions
|
@ -416,11 +416,11 @@ class OpenAIResponseObject(BaseModel):
|
||||||
created_at: int
|
created_at: int
|
||||||
error: Optional[OpenAIResponseError] = None
|
error: Optional[OpenAIResponseError] = None
|
||||||
id: str
|
id: str
|
||||||
incomplete_details: Optional[OpenAIResponseIncompleteDetails] = None # TODO: unimplemented
|
incomplete_details: Optional[OpenAIResponseIncompleteDetails] = None
|
||||||
instructions: Optional[str | list[str]] = None # TODO: unimplemented
|
instructions: Optional[str | list[str]] = None
|
||||||
max_output_tokens: Optional[int] = None # TODO: unimplemented
|
max_output_tokens: Optional[int] = None
|
||||||
max_tool_calls: Optional[int] = None # TODO: unimplemented
|
max_tool_calls: Optional[int] = None
|
||||||
metadata: Optional[dict[str, str]] = None # TODO: unimplemented
|
metadata: Optional[dict[str, str]] = None
|
||||||
model: str
|
model: str
|
||||||
object: Literal["response"] = "response"
|
object: Literal["response"] = "response"
|
||||||
output: list[OpenAIResponseOutput]
|
output: list[OpenAIResponseOutput]
|
||||||
|
@ -430,13 +430,13 @@ class OpenAIResponseObject(BaseModel):
|
||||||
prompt_cache_key: Optional[str] = None
|
prompt_cache_key: Optional[str] = None
|
||||||
reasoning: Optional[OpenAIResponseReasoning] = None
|
reasoning: Optional[OpenAIResponseReasoning] = None
|
||||||
safety_identifier: Optional[str] = None
|
safety_identifier: Optional[str] = None
|
||||||
service_tier: Optional[str] = None # TODO: unimplemented
|
service_tier: Optional[str] = None
|
||||||
status: str
|
status: str
|
||||||
temperature: float | None = None
|
temperature: float | None = None
|
||||||
text: Optional[OpenAIResponseText] = None
|
text: Optional[OpenAIResponseText] = None
|
||||||
tool_choice: Optional[OpenAIResponsesToolChoice] = None # TODO: unimplemented
|
tool_choice: Optional[OpenAIResponsesToolChoice] = None
|
||||||
tools: Optional[list[OpenAIResponsesTool]] = None # TODO: unimplemented
|
tools: Optional[list[OpenAIResponsesTool]] = None
|
||||||
top_logprobs: Optional[int] = None # TODO: unimplemented
|
top_logprobs: Optional[int] = None
|
||||||
top_p: Optional[float] = None
|
top_p: Optional[float] = None
|
||||||
user: Optional[str] = None # Deprecated: This field is being replaced by safety_identifier and prompt_cache_key
|
user: Optional[str] = None # Deprecated: This field is being replaced by safety_identifier and prompt_cache_key
|
||||||
truncation: Optional[str] = None
|
truncation: Optional[str] = None
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue