From 23f1980f9cb4390295254b3a43fc73a7eaacb2bd Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Wed, 22 Jan 2025 18:31:59 -0800 Subject: [PATCH] Fix meta-reference GPU implementation for inference --- .../providers/inline/inference/meta_reference/parallel_utils.py | 2 +- tests/client-sdk/inference/test_inference.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/llama_stack/providers/inline/inference/meta_reference/parallel_utils.py b/llama_stack/providers/inline/inference/meta_reference/parallel_utils.py index 36720612c..ced712257 100644 --- a/llama_stack/providers/inline/inference/meta_reference/parallel_utils.py +++ b/llama_stack/providers/inline/inference/meta_reference/parallel_utils.py @@ -357,8 +357,8 @@ class ModelParallelProcessGroup: assert not self.running, "inference already running" self.running = True - self.request_socket.send(encode_msg(TaskRequest(task=req))) try: + self.request_socket.send(encode_msg(TaskRequest(task=req))) while True: obj_json = self.request_socket.recv() obj = parse_message(obj_json) diff --git a/tests/client-sdk/inference/test_inference.py b/tests/client-sdk/inference/test_inference.py index 01bbd7dc0..8ca11521c 100644 --- a/tests/client-sdk/inference/test_inference.py +++ b/tests/client-sdk/inference/test_inference.py @@ -54,7 +54,7 @@ def base64_image_url(): with open(image_path, "rb") as image_file: # Convert the image to base64 base64_string = base64.b64encode(image_file.read()).decode("utf-8") - base64_url = f"data:image;base64,{base64_string}" + base64_url = f"data:image/png;base64,{base64_string}" return base64_url