diff --git a/llama_stack/providers/inline/agents/meta_reference/rag/context_retriever.py b/llama_stack/providers/inline/agents/meta_reference/rag/context_retriever.py index 1dbe7a91c..7b5c8b4b0 100644 --- a/llama_stack/providers/inline/agents/meta_reference/rag/context_retriever.py +++ b/llama_stack/providers/inline/agents/meta_reference/rag/context_retriever.py @@ -64,7 +64,7 @@ async def llm_rag_query_generator( model = config.model message = UserMessage(content=content) response = await inference_api.chat_completion( - model=model, + model_id=model, messages=[message], stream=False, )