diff --git a/llama_stack/providers/remote/inference/gemini/models.py b/llama_stack/providers/remote/inference/gemini/models.py index 1d7b47315..ef1cf339f 100644 --- a/llama_stack/providers/remote/inference/gemini/models.py +++ b/llama_stack/providers/remote/inference/gemini/models.py @@ -12,6 +12,9 @@ from llama_stack.providers.utils.inference.model_registry import ( LLM_MODEL_IDS = [ "gemini/gemini-1.5-flash", "gemini/gemini-1.5-pro", + "gemini/gemini-2.0-flash", + "gemini/gemini-2.5-flash", + "gemini/gemini-2.5-pro", ] diff --git a/llama_stack/templates/starter/run.yaml b/llama_stack/templates/starter/run.yaml index 319ababe5..960e96d01 100644 --- a/llama_stack/templates/starter/run.yaml +++ b/llama_stack/templates/starter/run.yaml @@ -731,6 +731,21 @@ models: provider_id: gemini provider_model_id: gemini/gemini-1.5-pro model_type: llm +- metadata: {} + model_id: gemini/gemini-2.0-flash + provider_id: gemini + provider_model_id: gemini/gemini-2.0-flash + model_type: llm +- metadata: {} + model_id: gemini/gemini-2.5-flash + provider_id: gemini + provider_model_id: gemini/gemini-2.5-flash + model_type: llm +- metadata: {} + model_id: gemini/gemini-2.5-pro + provider_id: gemini + provider_model_id: gemini/gemini-2.5-pro + model_type: llm - metadata: embedding_dimension: 768 context_length: 2048