mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
Fix Meta reference GPU implementation (#663)
By performing in-place mutations, we lost. Never in life do that.
This commit is contained in:
parent
f19eb8eee3
commit
540fc4d717
2 changed files with 15 additions and 7 deletions
|
@ -94,9 +94,14 @@ async def convert_request_to_raw(
|
|||
d = m.model_dump()
|
||||
d["content"] = content
|
||||
messages.append(RawMessage(**d))
|
||||
request.messages = messages
|
||||
|
||||
d = request.model_dump()
|
||||
d["messages"] = messages
|
||||
request = ChatCompletionRequestWithRawContent(**d)
|
||||
else:
|
||||
request.content = await interleaved_content_convert_to_raw(request.content)
|
||||
d = request.model_dump()
|
||||
d["content"] = await interleaved_content_convert_to_raw(request.content)
|
||||
request = CompletionRequestWithRawContent(**d)
|
||||
|
||||
return request
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue