diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 6f40ee1ee..59d787c8a 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -1988,7 +1988,7 @@ "supports_vision": true, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, - "gemini-experimental": { + "gemini-pro-experimental": { "max_tokens": 8192, "max_input_tokens": 1000000, "max_output_tokens": 8192, @@ -1998,7 +1998,19 @@ "mode": "chat", "supports_function_calling": false, "supports_tool_choice": true, - "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/gemini-experimental" + }, + "gemini-flash-experimental": { + "max_tokens": 8192, + "max_input_tokens": 1000000, + "max_output_tokens": 8192, + "input_cost_per_token": 0, + "output_cost_per_token": 0, + "litellm_provider": "vertex_ai-language-models", + "mode": "chat", + "supports_function_calling": false, + "supports_tool_choice": true, + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/gemini-experimental" }, "gemini-pro-vision": { "max_tokens": 2048, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 6f40ee1ee..ee22ecbe8 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -1988,7 +1988,7 @@ "supports_vision": true, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, - "gemini-experimental": { + "gemini-pro-experimental": { "max_tokens": 8192, "max_input_tokens": 1000000, "max_output_tokens": 8192, @@ -1998,7 +1998,19 @@ "mode": "chat", "supports_function_calling": false, "supports_tool_choice": true, - "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/gemini-experimental" + }, + "gemini-pro-flash": { + "max_tokens": 8192, + "max_input_tokens": 1000000, + "max_output_tokens": 8192, + "input_cost_per_token": 0, + "output_cost_per_token": 0, + "litellm_provider": "vertex_ai-language-models", + "mode": "chat", + "supports_function_calling": false, + "supports_tool_choice": true, + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/gemini-experimental" }, "gemini-pro-vision": { "max_tokens": 2048,