mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-18 23:28:53 +00:00
rebase and fix some small breakage due to model -> model_id fix
This commit is contained in:
parent
22aedd0277
commit
1cb42d3060
6 changed files with 20 additions and 11 deletions
|
@ -234,7 +234,7 @@ class LlamaGuardShield:
|
|||
# TODO: llama-stack inference protocol has issues with non-streaming inference code
|
||||
content = ""
|
||||
async for chunk in await self.inference_api.chat_completion(
|
||||
model=self.model,
|
||||
model_id=self.model,
|
||||
messages=[shield_input_message],
|
||||
stream=True,
|
||||
):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue