mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-08 21:04:39 +00:00
Merge 1e891489a8
into sapling-pr-archive-ehhuang
This commit is contained in:
commit
5025e02d81
13 changed files with 170 additions and 4 deletions
26
docs/static/deprecated-llama-stack-spec.html
vendored
26
docs/static/deprecated-llama-stack-spec.html
vendored
|
@ -7372,12 +7372,36 @@
|
|||
"user": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The user to use."
|
||||
},
|
||||
"kwargs": {
|
||||
"oneOf": [
|
||||
{
|
||||
"type": "null"
|
||||
},
|
||||
{
|
||||
"type": "boolean"
|
||||
},
|
||||
{
|
||||
"type": "number"
|
||||
},
|
||||
{
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"type": "array"
|
||||
},
|
||||
{
|
||||
"type": "object"
|
||||
}
|
||||
],
|
||||
"description": "(Optional) Additional provider-specific parameters to pass through as extra_body (e.g., chat_template_kwargs for vLLM)."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"model",
|
||||
"messages"
|
||||
"messages",
|
||||
"kwargs"
|
||||
],
|
||||
"title": "OpenaiChatCompletionRequest"
|
||||
},
|
||||
|
|
12
docs/static/deprecated-llama-stack-spec.yaml
vendored
12
docs/static/deprecated-llama-stack-spec.yaml
vendored
|
@ -5434,10 +5434,22 @@ components:
|
|||
user:
|
||||
type: string
|
||||
description: (Optional) The user to use.
|
||||
kwargs:
|
||||
oneOf:
|
||||
- type: 'null'
|
||||
- type: boolean
|
||||
- type: number
|
||||
- type: string
|
||||
- type: array
|
||||
- type: object
|
||||
description: >-
|
||||
(Optional) Additional provider-specific parameters to pass through as
|
||||
extra_body (e.g., chat_template_kwargs for vLLM).
|
||||
additionalProperties: false
|
||||
required:
|
||||
- model
|
||||
- messages
|
||||
- kwargs
|
||||
title: OpenaiChatCompletionRequest
|
||||
OpenAIChatCompletion:
|
||||
type: object
|
||||
|
|
26
docs/static/llama-stack-spec.html
vendored
26
docs/static/llama-stack-spec.html
vendored
|
@ -5167,12 +5167,36 @@
|
|||
"user": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The user to use."
|
||||
},
|
||||
"kwargs": {
|
||||
"oneOf": [
|
||||
{
|
||||
"type": "null"
|
||||
},
|
||||
{
|
||||
"type": "boolean"
|
||||
},
|
||||
{
|
||||
"type": "number"
|
||||
},
|
||||
{
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"type": "array"
|
||||
},
|
||||
{
|
||||
"type": "object"
|
||||
}
|
||||
],
|
||||
"description": "(Optional) Additional provider-specific parameters to pass through as extra_body (e.g., chat_template_kwargs for vLLM)."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"model",
|
||||
"messages"
|
||||
"messages",
|
||||
"kwargs"
|
||||
],
|
||||
"title": "OpenaiChatCompletionRequest"
|
||||
},
|
||||
|
|
12
docs/static/llama-stack-spec.yaml
vendored
12
docs/static/llama-stack-spec.yaml
vendored
|
@ -3888,10 +3888,22 @@ components:
|
|||
user:
|
||||
type: string
|
||||
description: (Optional) The user to use.
|
||||
kwargs:
|
||||
oneOf:
|
||||
- type: 'null'
|
||||
- type: boolean
|
||||
- type: number
|
||||
- type: string
|
||||
- type: array
|
||||
- type: object
|
||||
description: >-
|
||||
(Optional) Additional provider-specific parameters to pass through as
|
||||
extra_body (e.g., chat_template_kwargs for vLLM).
|
||||
additionalProperties: false
|
||||
required:
|
||||
- model
|
||||
- messages
|
||||
- kwargs
|
||||
title: OpenaiChatCompletionRequest
|
||||
OpenAIChatCompletion:
|
||||
type: object
|
||||
|
|
26
docs/static/stainless-llama-stack-spec.html
vendored
26
docs/static/stainless-llama-stack-spec.html
vendored
|
@ -6839,12 +6839,36 @@
|
|||
"user": {
|
||||
"type": "string",
|
||||
"description": "(Optional) The user to use."
|
||||
},
|
||||
"kwargs": {
|
||||
"oneOf": [
|
||||
{
|
||||
"type": "null"
|
||||
},
|
||||
{
|
||||
"type": "boolean"
|
||||
},
|
||||
{
|
||||
"type": "number"
|
||||
},
|
||||
{
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"type": "array"
|
||||
},
|
||||
{
|
||||
"type": "object"
|
||||
}
|
||||
],
|
||||
"description": "(Optional) Additional provider-specific parameters to pass through as extra_body (e.g., chat_template_kwargs for vLLM)."
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
"required": [
|
||||
"model",
|
||||
"messages"
|
||||
"messages",
|
||||
"kwargs"
|
||||
],
|
||||
"title": "OpenaiChatCompletionRequest"
|
||||
},
|
||||
|
|
12
docs/static/stainless-llama-stack-spec.yaml
vendored
12
docs/static/stainless-llama-stack-spec.yaml
vendored
|
@ -5101,10 +5101,22 @@ components:
|
|||
user:
|
||||
type: string
|
||||
description: (Optional) The user to use.
|
||||
kwargs:
|
||||
oneOf:
|
||||
- type: 'null'
|
||||
- type: boolean
|
||||
- type: number
|
||||
- type: string
|
||||
- type: array
|
||||
- type: object
|
||||
description: >-
|
||||
(Optional) Additional provider-specific parameters to pass through as
|
||||
extra_body (e.g., chat_template_kwargs for vLLM).
|
||||
additionalProperties: false
|
||||
required:
|
||||
- model
|
||||
- messages
|
||||
- kwargs
|
||||
title: OpenaiChatCompletionRequest
|
||||
OpenAIChatCompletion:
|
||||
type: object
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue