mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-07 12:47:37 +00:00
Further bug fixes
This commit is contained in:
parent
9e16b0948b
commit
e5a7001874
3 changed files with 30 additions and 18 deletions
|
@ -248,7 +248,7 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
self,
|
||||
turn_id: str,
|
||||
messages: List[Message],
|
||||
shields: List[ShieldDefinition],
|
||||
shields: List[str],
|
||||
touchpoint: str,
|
||||
) -> AsyncGenerator:
|
||||
if len(shields) == 0:
|
||||
|
@ -608,7 +608,6 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
else:
|
||||
return True
|
||||
|
||||
print(f"{enabled_tools=}")
|
||||
return AgentTool.memory.value in enabled_tools
|
||||
|
||||
def _memory_tool_definition(self) -> Optional[MemoryToolDefinition]:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue