diff --git a/llama_stack/apis/agents/agents.py b/llama_stack/apis/agents/agents.py index ba383d1f7..ea49c3479 100644 --- a/llama_stack/apis/agents/agents.py +++ b/llama_stack/apis/agents/agents.py @@ -26,6 +26,7 @@ from llama_stack.apis.common.content_types import ContentDelta, InterleavedConte from llama_stack.apis.inference import ( CompletionMessage, ResponseFormat, + ResponseOutputParser, SamplingParams, ToolCall, ToolChoice, @@ -165,7 +166,7 @@ class AgentConfig(AgentConfigCommon): instructions: str enable_session_persistence: bool response_format: Optional[ResponseFormat] = None - output_parser: Optional[OutputParser] = Field(default=OutputParser.auto) + response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default) class AgentConfigOverridablePerTurn(AgentConfigCommon): diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py index 637b39297..4b1ee82d9 100644 --- a/llama_stack/apis/inference/inference.py +++ b/llama_stack/apis/inference/inference.py @@ -320,7 +320,7 @@ class ChatCompletionRequest(BaseModel): response_format: Optional[ResponseFormat] = None stream: Optional[bool] = False logprobs: Optional[LogProbConfig] = None - response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.auto) + response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default) @json_schema_type diff --git a/llama_stack/providers/utils/inference/prompt_adapter.py b/llama_stack/providers/utils/inference/prompt_adapter.py index 8156c73d3..49c6ac7a9 100644 --- a/llama_stack/providers/utils/inference/prompt_adapter.py +++ b/llama_stack/providers/utils/inference/prompt_adapter.py @@ -375,7 +375,6 @@ def augment_messages_for_tools_llama_3_1( def augment_messages_for_tools_llama_3_2( request: ChatCompletionRequest, ) -> List[Message]: - breakpoint() assert request.tool_choice == ToolChoice.auto, "Only `ToolChoice.auto` supported" existing_messages = request.messages