mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-07 02:58:21 +00:00
api change
This commit is contained in:
parent
4d07460398
commit
662f171e88
3 changed files with 3 additions and 3 deletions
|
@ -26,6 +26,7 @@ from llama_stack.apis.common.content_types import ContentDelta, InterleavedConte
|
|||
from llama_stack.apis.inference import (
|
||||
CompletionMessage,
|
||||
ResponseFormat,
|
||||
ResponseOutputParser,
|
||||
SamplingParams,
|
||||
ToolCall,
|
||||
ToolChoice,
|
||||
|
@ -165,7 +166,7 @@ class AgentConfig(AgentConfigCommon):
|
|||
instructions: str
|
||||
enable_session_persistence: bool
|
||||
response_format: Optional[ResponseFormat] = None
|
||||
output_parser: Optional[OutputParser] = Field(default=OutputParser.auto)
|
||||
response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default)
|
||||
|
||||
|
||||
class AgentConfigOverridablePerTurn(AgentConfigCommon):
|
||||
|
|
|
@ -320,7 +320,7 @@ class ChatCompletionRequest(BaseModel):
|
|||
response_format: Optional[ResponseFormat] = None
|
||||
stream: Optional[bool] = False
|
||||
logprobs: Optional[LogProbConfig] = None
|
||||
response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.auto)
|
||||
response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default)
|
||||
|
||||
|
||||
@json_schema_type
|
||||
|
|
|
@ -375,7 +375,6 @@ def augment_messages_for_tools_llama_3_1(
|
|||
def augment_messages_for_tools_llama_3_2(
|
||||
request: ChatCompletionRequest,
|
||||
) -> List[Message]:
|
||||
breakpoint()
|
||||
assert request.tool_choice == ToolChoice.auto, "Only `ToolChoice.auto` supported"
|
||||
|
||||
existing_messages = request.messages
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue