mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 19:04:19 +00:00
Fix meta-reference GPU implementation for inference
This commit is contained in:
parent
f4b0f2af8b
commit
23f1980f9c
2 changed files with 2 additions and 2 deletions
|
@ -54,7 +54,7 @@ def base64_image_url():
|
|||
with open(image_path, "rb") as image_file:
|
||||
# Convert the image to base64
|
||||
base64_string = base64.b64encode(image_file.read()).decode("utf-8")
|
||||
base64_url = f"data:image;base64,{base64_string}"
|
||||
base64_url = f"data:image/png;base64,{base64_string}"
|
||||
return base64_url
|
||||
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue