Update prompt adapter

This commit is contained in:
Aidan Do 2024-12-19 22:09:23 +11:00
parent 94645dd5f6
commit 4331796692

View file

@ -271,11 +271,15 @@ def chat_completion_request_to_messages(
log.error(f"Unsupported inference model? {model.descriptor()}")
return request.messages
if model.model_family == ModelFamily.llama3_1 or (
model.model_family == ModelFamily.llama3_2
and is_multimodal(model.core_model_id)
if (
model.model_family == ModelFamily.llama3_1
or (
model.model_family == ModelFamily.llama3_2
and is_multimodal(model.core_model_id)
)
or model.model_family == ModelFamily.llama3_3
):
# llama3.1 and llama3.2 multimodal models follow the same tool prompt format
# llama3.1, llama3.2 multimodal and llama3.3 models follow the same tool prompt format
messages = augment_messages_for_tools_llama_3_1(request)
elif model.model_family == ModelFamily.llama3_2:
messages = augment_messages_for_tools_llama_3_2(request)