diff --git a/llama_stack/providers/utils/inference/prompt_adapter.py b/llama_stack/providers/utils/inference/prompt_adapter.py index 9f034e801..66888ac37 100644 --- a/llama_stack/providers/utils/inference/prompt_adapter.py +++ b/llama_stack/providers/utils/inference/prompt_adapter.py @@ -271,11 +271,15 @@ def chat_completion_request_to_messages( log.error(f"Unsupported inference model? {model.descriptor()}") return request.messages - if model.model_family == ModelFamily.llama3_1 or ( - model.model_family == ModelFamily.llama3_2 - and is_multimodal(model.core_model_id) + if ( + model.model_family == ModelFamily.llama3_1 + or ( + model.model_family == ModelFamily.llama3_2 + and is_multimodal(model.core_model_id) + ) + or model.model_family == ModelFamily.llama3_3 ): - # llama3.1 and llama3.2 multimodal models follow the same tool prompt format + # llama3.1, llama3.2 multimodal and llama3.3 models follow the same tool prompt format messages = augment_messages_for_tools_llama_3_1(request) elif model.model_family == ModelFamily.llama3_2: messages = augment_messages_for_tools_llama_3_2(request)