From 1a5cfd1b6fece6b53cc708f49bf4b145c400b9eb Mon Sep 17 00:00:00 2001 From: Neil Mehta Date: Fri, 14 Mar 2025 15:51:12 -0400 Subject: [PATCH] Fix stream generate --- .../providers/remote/inference/lmstudio/_client.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/llama_stack/providers/remote/inference/lmstudio/_client.py b/llama_stack/providers/remote/inference/lmstudio/_client.py index 84bbd946f..fc7d626b1 100644 --- a/llama_stack/providers/remote/inference/lmstudio/_client.py +++ b/llama_stack/providers/remote/inference/lmstudio/_client.py @@ -116,12 +116,12 @@ class LMStudioClient: delta=TextDelta(text=chunk.content), ) ) - yield ChatCompletionResponseStreamChunk( - event=ChatCompletionResponseEvent( - event_type=ChatCompletionResponseEventType.complete, - delta=TextDelta(text=""), - ) + yield ChatCompletionResponseStreamChunk( + event=ChatCompletionResponseEvent( + event_type=ChatCompletionResponseEventType.complete, + delta=TextDelta(text=""), ) + ) return stream_generator() else: