diff --git a/llama_stack/providers/remote/inference/fireworks/config.py b/llama_stack/providers/remote/inference/fireworks/config.py index e69926942..979e8455a 100644 --- a/llama_stack/providers/remote/inference/fireworks/config.py +++ b/llama_stack/providers/remote/inference/fireworks/config.py @@ -22,7 +22,7 @@ class FireworksImplConfig(BaseModel): ) @classmethod - def sample_run_config(cls) -> Dict[str, Any]: + def sample_run_config(cls, __distro_dir__: str) -> Dict[str, Any]: return { "url": "https://api.fireworks.ai/inference/v1", "api_key": "${env.FIREWORKS_API_KEY}", diff --git a/llama_stack/providers/remote/inference/fireworks/fireworks.py b/llama_stack/providers/remote/inference/fireworks/fireworks.py index d9ef57b15..975ec4893 100644 --- a/llama_stack/providers/remote/inference/fireworks/fireworks.py +++ b/llama_stack/providers/remote/inference/fireworks/fireworks.py @@ -65,6 +65,10 @@ MODEL_ALIASES = [ "fireworks/llama-v3p2-90b-vision-instruct", CoreModelId.llama3_2_90b_vision_instruct.value, ), + build_model_alias( + "fireworks/llama-v3p3-70b-instruct", + CoreModelId.llama3_3_70b_instruct.value, + ), build_model_alias( "fireworks/llama-guard-3-8b", CoreModelId.llama_guard_3_8b.value, diff --git a/llama_stack/providers/utils/inference/prompt_adapter.py b/llama_stack/providers/utils/inference/prompt_adapter.py index 82fcefe54..f7d2cd84e 100644 --- a/llama_stack/providers/utils/inference/prompt_adapter.py +++ b/llama_stack/providers/utils/inference/prompt_adapter.py @@ -282,7 +282,8 @@ def chat_completion_request_to_messages( ): # llama3.1 and llama3.2 multimodal models follow the same tool prompt format messages = augment_messages_for_tools_llama_3_1(request) - elif model.model_family == ModelFamily.llama3_2: + elif model.model_family in (ModelFamily.llama3_2, ModelFamily.llama3_3): + # llama3.2 and llama3.3 models follow the same tool prompt format messages = augment_messages_for_tools_llama_3_2(request) else: messages = request.messages diff --git a/llama_stack/templates/fireworks/run.yaml b/llama_stack/templates/fireworks/run.yaml index cb31b4678..99f155a4a 100644 --- a/llama_stack/templates/fireworks/run.yaml +++ b/llama_stack/templates/fireworks/run.yaml @@ -110,6 +110,11 @@ models: provider_id: fireworks provider_model_id: fireworks/llama-v3p2-90b-vision-instruct model_type: llm +- metadata: {} + model_id: meta-llama/Llama-3.3-70B-Instruct + provider_id: fireworks + provider_model_id: fireworks/llama-v3p3-70b-instruct + model_type: llm - metadata: {} model_id: meta-llama/Llama-Guard-3-8B provider_id: fireworks