From 1677d6bffdf0a002abe3b827c460df4930ee83c8 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Vlastimil=20Eli=C3=A1=C5=A1?= Date: Fri, 8 Aug 2025 22:48:15 +0200 Subject: [PATCH] feat: Flash-Lite 2.0 and 2.5 models added to Gemini inference provider (#3058) PR adds Flash-Lite 2.0 and 2.5 models to the Gemini inference provider Closes #3046 ## Test Plan I was not able to locate any existing test for this provider, so I performed manual testing. But the change is really trivial and straightforward. --- llama_stack/providers/remote/inference/gemini/models.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/llama_stack/providers/remote/inference/gemini/models.py b/llama_stack/providers/remote/inference/gemini/models.py index 6fda35e0f..bd696b0ac 100644 --- a/llama_stack/providers/remote/inference/gemini/models.py +++ b/llama_stack/providers/remote/inference/gemini/models.py @@ -13,7 +13,9 @@ LLM_MODEL_IDS = [ "gemini-1.5-flash", "gemini-1.5-pro", "gemini-2.0-flash", + "gemini-2.0-flash-lite", "gemini-2.5-flash", + "gemini-2.5-flash-lite", "gemini-2.5-pro", ]