From c20f342cdee26bfc66d13705970959d6cf2fd674 Mon Sep 17 00:00:00 2001 From: Aidan Do Date: Fri, 20 Dec 2024 08:58:45 +1100 Subject: [PATCH] Move to 3_2 format --- .../providers/utils/inference/prompt_adapter.py | 12 ++++-------- 1 file changed, 4 insertions(+), 8 deletions(-) diff --git a/llama_stack/providers/utils/inference/prompt_adapter.py b/llama_stack/providers/utils/inference/prompt_adapter.py index 66888ac37..ae030134e 100644 --- a/llama_stack/providers/utils/inference/prompt_adapter.py +++ b/llama_stack/providers/utils/inference/prompt_adapter.py @@ -271,17 +271,13 @@ def chat_completion_request_to_messages( log.error(f"Unsupported inference model? {model.descriptor()}") return request.messages - if ( - model.model_family == ModelFamily.llama3_1 - or ( - model.model_family == ModelFamily.llama3_2 - and is_multimodal(model.core_model_id) - ) - or model.model_family == ModelFamily.llama3_3 + if model.model_family == ModelFamily.llama3_1 or ( + model.model_family == ModelFamily.llama3_2 + and is_multimodal(model.core_model_id) ): # llama3.1, llama3.2 multimodal and llama3.3 models follow the same tool prompt format messages = augment_messages_for_tools_llama_3_1(request) - elif model.model_family == ModelFamily.llama3_2: + elif model.model_family in (ModelFamily.llama3_2, ModelFamily.llama3_3): messages = augment_messages_for_tools_llama_3_2(request) else: messages = request.messages