forked from phoenix-oss/llama-stack-mirror
fix context retriever (#75)
This commit is contained in:
parent
055770a791
commit
e6fdb9df29
1 changed files with 3 additions and 5 deletions
|
@ -63,11 +63,9 @@ async def llm_rag_query_generator(
|
|||
model = config.model
|
||||
message = UserMessage(content=content)
|
||||
response = inference_api.chat_completion(
|
||||
ChatCompletionRequest(
|
||||
model=model,
|
||||
messages=[message],
|
||||
stream=False,
|
||||
)
|
||||
model=model,
|
||||
messages=[message],
|
||||
stream=False,
|
||||
)
|
||||
|
||||
async for chunk in response:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue