mirror of
https://github.com/meta-llama/llama-stack.git
synced 2026-01-02 19:00:02 +00:00
fix again vllm
This commit is contained in:
parent
3e7496e835
commit
9a0fa89e9a
2 changed files with 3 additions and 2 deletions
|
|
@ -177,7 +177,8 @@ class VLLMInferenceAdapter(Inference, ModelsProtocolPrivate):
|
|||
if isinstance(request, ChatCompletionRequest):
|
||||
if media_present:
|
||||
input_dict["messages"] = [
|
||||
await convert_message_to_openai_dict(m) for m in request.messages
|
||||
await convert_message_to_openai_dict(m, download=True)
|
||||
for m in request.messages
|
||||
]
|
||||
else:
|
||||
input_dict["prompt"] = await chat_completion_request_to_prompt(
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue