mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-07 02:58:21 +00:00
api change
This commit is contained in:
parent
4d07460398
commit
662f171e88
3 changed files with 3 additions and 3 deletions
|
@ -26,6 +26,7 @@ from llama_stack.apis.common.content_types import ContentDelta, InterleavedConte
|
||||||
from llama_stack.apis.inference import (
|
from llama_stack.apis.inference import (
|
||||||
CompletionMessage,
|
CompletionMessage,
|
||||||
ResponseFormat,
|
ResponseFormat,
|
||||||
|
ResponseOutputParser,
|
||||||
SamplingParams,
|
SamplingParams,
|
||||||
ToolCall,
|
ToolCall,
|
||||||
ToolChoice,
|
ToolChoice,
|
||||||
|
@ -165,7 +166,7 @@ class AgentConfig(AgentConfigCommon):
|
||||||
instructions: str
|
instructions: str
|
||||||
enable_session_persistence: bool
|
enable_session_persistence: bool
|
||||||
response_format: Optional[ResponseFormat] = None
|
response_format: Optional[ResponseFormat] = None
|
||||||
output_parser: Optional[OutputParser] = Field(default=OutputParser.auto)
|
response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default)
|
||||||
|
|
||||||
|
|
||||||
class AgentConfigOverridablePerTurn(AgentConfigCommon):
|
class AgentConfigOverridablePerTurn(AgentConfigCommon):
|
||||||
|
|
|
@ -320,7 +320,7 @@ class ChatCompletionRequest(BaseModel):
|
||||||
response_format: Optional[ResponseFormat] = None
|
response_format: Optional[ResponseFormat] = None
|
||||||
stream: Optional[bool] = False
|
stream: Optional[bool] = False
|
||||||
logprobs: Optional[LogProbConfig] = None
|
logprobs: Optional[LogProbConfig] = None
|
||||||
response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.auto)
|
response_output_parser: Optional[ResponseOutputParser] = Field(default=ResponseOutputParser.default)
|
||||||
|
|
||||||
|
|
||||||
@json_schema_type
|
@json_schema_type
|
||||||
|
|
|
@ -375,7 +375,6 @@ def augment_messages_for_tools_llama_3_1(
|
||||||
def augment_messages_for_tools_llama_3_2(
|
def augment_messages_for_tools_llama_3_2(
|
||||||
request: ChatCompletionRequest,
|
request: ChatCompletionRequest,
|
||||||
) -> List[Message]:
|
) -> List[Message]:
|
||||||
breakpoint()
|
|
||||||
assert request.tool_choice == ToolChoice.auto, "Only `ToolChoice.auto` supported"
|
assert request.tool_choice == ToolChoice.auto, "Only `ToolChoice.auto` supported"
|
||||||
|
|
||||||
existing_messages = request.messages
|
existing_messages = request.messages
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue