From 237cc26ac69e2b0930236ec6e12be47cda2e71e0 Mon Sep 17 00:00:00 2001 From: Aidan Do Date: Fri, 20 Dec 2024 09:00:52 +1100 Subject: [PATCH] adapter --- llama_stack/providers/utils/inference/prompt_adapter.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/llama_stack/providers/utils/inference/prompt_adapter.py b/llama_stack/providers/utils/inference/prompt_adapter.py index ae030134e..2ed43c2f4 100644 --- a/llama_stack/providers/utils/inference/prompt_adapter.py +++ b/llama_stack/providers/utils/inference/prompt_adapter.py @@ -275,9 +275,10 @@ def chat_completion_request_to_messages( model.model_family == ModelFamily.llama3_2 and is_multimodal(model.core_model_id) ): - # llama3.1, llama3.2 multimodal and llama3.3 models follow the same tool prompt format + # llama3.1 and llama3.2 multimodal models follow the same tool prompt format messages = augment_messages_for_tools_llama_3_1(request) elif model.model_family in (ModelFamily.llama3_2, ModelFamily.llama3_3): + # llama3.2 and llama3.3 models follow the same tool prompt format messages = augment_messages_for_tools_llama_3_2(request) else: messages = request.messages