mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
fix: Remove deprecated user param in OpenAIResponseObject (#3596)
# What does this PR do? Just removing the deprecated User param in `OpenAIResponseObject` Closing https://github.com/llamastack/llama-stack/issues/3482 ## Test Plan CI
This commit is contained in:
parent
e9eb004bf8
commit
455579a88e
5 changed files with 0 additions and 20 deletions
8
docs/static/llama-stack-spec.html
vendored
8
docs/static/llama-stack-spec.html
vendored
|
@ -9456,10 +9456,6 @@
|
|||
"truncation": {
|
||||
"type": "string",
|
||||
"description": "(Optional) Truncation strategy applied to the response"
|
||||
},
|
||||
"user": {
|
||||
"type": "string",
|
||||
"description": "(Optional) User identifier associated with the request"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
|
@ -13594,10 +13590,6 @@
|
|||
"type": "string",
|
||||
"description": "(Optional) Truncation strategy applied to the response"
|
||||
},
|
||||
"user": {
|
||||
"type": "string",
|
||||
"description": "(Optional) User identifier associated with the request"
|
||||
},
|
||||
"input": {
|
||||
"type": "array",
|
||||
"items": {
|
||||
|
|
8
docs/static/llama-stack-spec.yaml
vendored
8
docs/static/llama-stack-spec.yaml
vendored
|
@ -6884,10 +6884,6 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) Truncation strategy applied to the response
|
||||
user:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) User identifier associated with the request
|
||||
additionalProperties: false
|
||||
required:
|
||||
- created_at
|
||||
|
@ -10082,10 +10078,6 @@ components:
|
|||
type: string
|
||||
description: >-
|
||||
(Optional) Truncation strategy applied to the response
|
||||
user:
|
||||
type: string
|
||||
description: >-
|
||||
(Optional) User identifier associated with the request
|
||||
input:
|
||||
type: array
|
||||
items:
|
||||
|
|
|
@ -336,7 +336,6 @@ class OpenAIResponseObject(BaseModel):
|
|||
:param text: Text formatting configuration for the response
|
||||
:param top_p: (Optional) Nucleus sampling parameter used for generation
|
||||
:param truncation: (Optional) Truncation strategy applied to the response
|
||||
:param user: (Optional) User identifier associated with the request
|
||||
"""
|
||||
|
||||
created_at: int
|
||||
|
@ -354,7 +353,6 @@ class OpenAIResponseObject(BaseModel):
|
|||
text: OpenAIResponseText = OpenAIResponseText(format=OpenAIResponseTextFormat(type="text"))
|
||||
top_p: float | None = None
|
||||
truncation: str | None = None
|
||||
user: str | None = None
|
||||
|
||||
|
||||
@json_schema_type
|
||||
|
|
|
@ -41,7 +41,6 @@ export default function ResponseDetailPage() {
|
|||
temperature: responseData.temperature,
|
||||
top_p: responseData.top_p,
|
||||
truncation: responseData.truncation,
|
||||
user: responseData.user,
|
||||
};
|
||||
};
|
||||
|
||||
|
|
|
@ -43,7 +43,6 @@ const convertResponseListData = (
|
|||
temperature: responseData.temperature,
|
||||
top_p: responseData.top_p,
|
||||
truncation: responseData.truncation,
|
||||
user: responseData.user,
|
||||
};
|
||||
};
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue