mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 19:04:19 +00:00
Fix meta-reference GPU implementation for inference
This commit is contained in:
parent
f4b0f2af8b
commit
23f1980f9c
2 changed files with 2 additions and 2 deletions
|
@ -357,8 +357,8 @@ class ModelParallelProcessGroup:
|
|||
assert not self.running, "inference already running"
|
||||
|
||||
self.running = True
|
||||
self.request_socket.send(encode_msg(TaskRequest(task=req)))
|
||||
try:
|
||||
self.request_socket.send(encode_msg(TaskRequest(task=req)))
|
||||
while True:
|
||||
obj_json = self.request_socket.recv()
|
||||
obj = parse_message(obj_json)
|
||||
|
|
|
@ -54,7 +54,7 @@ def base64_image_url():
|
|||
with open(image_path, "rb") as image_file:
|
||||
# Convert the image to base64
|
||||
base64_string = base64.b64encode(image_file.read()).decode("utf-8")
|
||||
base64_url = f"data:image;base64,{base64_string}"
|
||||
base64_url = f"data:image/png;base64,{base64_string}"
|
||||
return base64_url
|
||||
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue