mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-03 01:03:59 +00:00
fix shields step
This commit is contained in:
parent
bf961f8aa5
commit
aab4084ee3
1 changed files with 0 additions and 2 deletions
|
@ -240,7 +240,6 @@ class ChatAgent(ShieldRunnerMixin):
|
||||||
# final boolean (to see whether an exception happened) and then explicitly testing for it.
|
# final boolean (to see whether an exception happened) and then explicitly testing for it.
|
||||||
|
|
||||||
if len(self.input_shields) > 0:
|
if len(self.input_shields) > 0:
|
||||||
print("input shields", self.input_shields)
|
|
||||||
async for res in self.run_multiple_shields_wrapper(
|
async for res in self.run_multiple_shields_wrapper(
|
||||||
turn_id, input_messages, self.input_shields, "user-input"
|
turn_id, input_messages, self.input_shields, "user-input"
|
||||||
):
|
):
|
||||||
|
@ -282,7 +281,6 @@ class ChatAgent(ShieldRunnerMixin):
|
||||||
shields: List[str],
|
shields: List[str],
|
||||||
touchpoint: str,
|
touchpoint: str,
|
||||||
) -> AsyncGenerator:
|
) -> AsyncGenerator:
|
||||||
print("!!! HI run_multiple_shields_wrapper")
|
|
||||||
with tracing.span("run_shields") as span:
|
with tracing.span("run_shields") as span:
|
||||||
span.set_attribute("input", [m.model_dump_json() for m in messages])
|
span.set_attribute("input", [m.model_dump_json() for m in messages])
|
||||||
if len(shields) == 0:
|
if len(shields) == 0:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue