From 42f0e919090e90379aebf0dedb291c70b517b9cf Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Tue, 4 Feb 2025 17:33:08 -0800 Subject: [PATCH] remove response parser --- llama_stack/apis/agents/agents.py | 2 -- llama_stack/apis/inference/inference.py | 2 -- 2 files changed, 4 deletions(-) diff --git a/llama_stack/apis/agents/agents.py b/llama_stack/apis/agents/agents.py index ea49c3479..50bea3d55 100644 --- a/llama_stack/apis/agents/agents.py +++ b/llama_stack/apis/agents/agents.py @@ -26,7 +26,6 @@ from llama_stack.apis.common.content_types import ContentDelta, InterleavedConte from llama_stack.apis.inference import ( CompletionMessage, ResponseFormat, - ResponseOutputParser, SamplingParams, ToolCall, ToolChoice, @@ -166,7 +165,6 @@ class AgentConfig(AgentConfigCommon): instructions: str enable_session_persistence: bool response_format: Optional[ResponseFormat] = None - response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default) class AgentConfigOverridablePerTurn(AgentConfigCommon): diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py index 4b1ee82d9..6398f74e8 100644 --- a/llama_stack/apis/inference/inference.py +++ b/llama_stack/apis/inference/inference.py @@ -19,7 +19,6 @@ from typing import ( from llama_models.llama3.api.datatypes import ( BuiltinTool, - ResponseOutputParser, SamplingParams, StopReason, ToolCall, @@ -320,7 +319,6 @@ class ChatCompletionRequest(BaseModel): response_format: Optional[ResponseFormat] = None stream: Optional[bool] = False logprobs: Optional[LogProbConfig] = None - response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default) @json_schema_type