mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 01:48:05 +00:00
Remove incremental_function_calling from method signature
Remove incremental_function_calling parameter from method.
This commit is contained in:
parent
07a2aa200c
commit
6ba51a0851
6 changed files with 62 additions and 62 deletions
|
|
@ -6723,12 +6723,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
title: Parallel Tool Calls
|
||||
type: boolean
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
description: >-
|
||||
(Optional) Whether to allow more than one function tool call generated
|
||||
per turn.
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -6838,14 +6836,12 @@ components:
|
|||
title: Input
|
||||
type: object
|
||||
required:
|
||||
- created_at
|
||||
- id
|
||||
- model
|
||||
- object
|
||||
- output
|
||||
- status
|
||||
- text
|
||||
- input
|
||||
- created_at
|
||||
- id
|
||||
- model
|
||||
- output
|
||||
- status
|
||||
- input
|
||||
title: OpenAIResponseObjectWithInput
|
||||
description: OpenAI response object extended with input context information.
|
||||
OpenAIResponseOutput:
|
||||
|
|
@ -7127,12 +7123,14 @@ components:
|
|||
- type: 'null'
|
||||
title: OpenAIResponsePrompt
|
||||
instructions:
|
||||
type: string
|
||||
anyOf:
|
||||
- type: string
|
||||
- type: 'null'
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -7259,12 +7257,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
description: >-
|
||||
(Optional) Whether to allow more than one function tool call generated
|
||||
per turn.
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -7336,11 +7332,11 @@ components:
|
|||
- type: 'null'
|
||||
type: object
|
||||
required:
|
||||
- created_at
|
||||
- id
|
||||
- model
|
||||
- output
|
||||
- status
|
||||
- created_at
|
||||
- id
|
||||
- model
|
||||
- output
|
||||
- status
|
||||
title: OpenAIResponseObject
|
||||
description: Complete OpenAI response object containing generation results and metadata.
|
||||
OpenAIResponseContentPartOutputText:
|
||||
|
|
|
|||
19
docs/static/deprecated-llama-stack-spec.yaml
vendored
19
docs/static/deprecated-llama-stack-spec.yaml
vendored
|
|
@ -3566,9 +3566,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
default: false
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -3968,6 +3969,11 @@ components:
|
|||
anyOf:
|
||||
- type: string
|
||||
- type: 'null'
|
||||
parallel_tool_calls:
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -4094,9 +4100,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
default: false
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
|
|||
14
docs/static/experimental-llama-stack-spec.yaml
vendored
14
docs/static/experimental-llama-stack-spec.yaml
vendored
|
|
@ -3263,9 +3263,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
default: false
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -3662,9 +3663,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
default: false
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
|
|||
22
docs/static/llama-stack-spec.yaml
vendored
22
docs/static/llama-stack-spec.yaml
vendored
|
|
@ -5744,12 +5744,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
description: >-
|
||||
(Optional) Whether to allow more than one function tool call generated
|
||||
per turn.
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -6146,12 +6144,14 @@ components:
|
|||
- type: 'null'
|
||||
title: OpenAIResponsePrompt
|
||||
instructions:
|
||||
type: string
|
||||
anyOf:
|
||||
- type: string
|
||||
- type: 'null'
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -6278,12 +6278,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
description: >-
|
||||
(Optional) Whether to allow more than one function tool call generated
|
||||
per turn.
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
|
|||
22
docs/static/stainless-llama-stack-spec.yaml
vendored
22
docs/static/stainless-llama-stack-spec.yaml
vendored
|
|
@ -6723,12 +6723,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
description: >-
|
||||
(Optional) Whether to allow more than one function tool call generated
|
||||
per turn.
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -7128,9 +7126,11 @@ components:
|
|||
anyOf:
|
||||
- type: string
|
||||
- type: 'null'
|
||||
type: string
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
@ -7257,12 +7257,10 @@ components:
|
|||
type: array
|
||||
title: Output
|
||||
parallel_tool_calls:
|
||||
type: boolean
|
||||
title: Parallel Tool Calls
|
||||
anyOf:
|
||||
- type: boolean
|
||||
- type: 'null'
|
||||
default: true
|
||||
description: >-
|
||||
(Optional) Whether to allow more than one function tool call generated
|
||||
per turn.
|
||||
previous_response_id:
|
||||
anyOf:
|
||||
- type: string
|
||||
|
|
|
|||
|
|
@ -1025,7 +1025,6 @@ class StreamingResponseOrchestrator:
|
|||
sequence_number=self.sequence_number,
|
||||
)
|
||||
|
||||
|
||||
async def _process_new_tools(
|
||||
self, tools: list[OpenAIResponseInputTool], output_messages: list[OpenAIResponseOutput]
|
||||
) -> AsyncIterator[OpenAIResponseObjectStream]:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue