From da47273dd1024ed337817ccc38f8ccedb13bda1f Mon Sep 17 00:00:00 2001 From: dltn <6599399+dltn@users.noreply.github.com> Date: Wed, 9 Oct 2024 11:56:02 -0700 Subject: [PATCH] second round of feedback --- .../impls/meta_reference/inference/parallel_utils.py | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/llama_stack/providers/impls/meta_reference/inference/parallel_utils.py b/llama_stack/providers/impls/meta_reference/inference/parallel_utils.py index 0fc883bec..c6eacc73c 100644 --- a/llama_stack/providers/impls/meta_reference/inference/parallel_utils.py +++ b/llama_stack/providers/impls/meta_reference/inference/parallel_utils.py @@ -56,28 +56,24 @@ class ReadyRequest(BaseModel): type: Literal[ProcessingMessageName.ready_request] = ( ProcessingMessageName.ready_request ) - message: str = "READY?" class ReadyResponse(BaseModel): type: Literal[ProcessingMessageName.ready_response] = ( ProcessingMessageName.ready_response ) - message: str = "YES READY" class EndSentinel(BaseModel): type: Literal[ProcessingMessageName.end_sentinel] = ( ProcessingMessageName.end_sentinel ) - message: str = "__end_sentinel__" class CancelSentinel(BaseModel): type: Literal[ProcessingMessageName.cancel_sentinel] = ( ProcessingMessageName.cancel_sentinel ) - message: str = "__cancel_sentinel__" class TaskRequest(BaseModel): @@ -143,7 +139,7 @@ def retrieve_requests(reply_socket_url: str): reply_socket.send_multipart([client_id, encode_msg(ready_response)]) break - def send_obj(obj: Union[ProcessingMessage, BaseModel]): + def send_obj(obj: ProcessingMessage): reply_socket.send_multipart([client_id, encode_msg(obj)]) while True: