From b0425c84b5be597d85e476e2c554161921ca4ad1 Mon Sep 17 00:00:00 2001 From: Hardik Shah Date: Tue, 18 Mar 2025 13:57:36 -0700 Subject: [PATCH] fix vllm bugs --- llama_stack/providers/remote/inference/vllm/vllm.py | 9 +++------ 1 file changed, 3 insertions(+), 6 deletions(-) diff --git a/llama_stack/providers/remote/inference/vllm/vllm.py b/llama_stack/providers/remote/inference/vllm/vllm.py index 8e494a846..6522820e0 100644 --- a/llama_stack/providers/remote/inference/vllm/vllm.py +++ b/llama_stack/providers/remote/inference/vllm/vllm.py @@ -89,16 +89,12 @@ def _convert_to_vllm_tool_calls_in_response( if not tool_calls: return [] - call_function_arguments = None - for call in tool_calls: - call_function_arguments = json.loads(call.function.arguments) - return [ ToolCall( call_id=call.id, tool_name=call.function.name, arguments=json.loads(call.function.arguments), - arguments_json=call_function_arguments, + arguments_json=call.function.arguments, ) for call in tool_calls ] @@ -182,7 +178,8 @@ async def _process_vllm_chat_completion_stream_response( tool_call=ToolCall( call_id=tool_call_buf.call_id, tool_name=tool_call_buf.tool_name, - arguments=args_str, + arguments=args, + arguments_json=args_str, ), parse_status=ToolCallParseStatus.succeeded, ),