diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 59d787c8ab..d1d1d602ca 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -1994,6 +1994,8 @@ "max_output_tokens": 8192, "input_cost_per_token": 0, "output_cost_per_token": 0, + "input_cost_per_character": 0, + "output_cost_per_character": 0, "litellm_provider": "vertex_ai-language-models", "mode": "chat", "supports_function_calling": false, @@ -2006,6 +2008,8 @@ "max_output_tokens": 8192, "input_cost_per_token": 0, "output_cost_per_token": 0, + "input_cost_per_character": 0, + "output_cost_per_character": 0, "litellm_provider": "vertex_ai-language-models", "mode": "chat", "supports_function_calling": false, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index ee22ecbe8d..d88b28ba1f 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -1994,6 +1994,8 @@ "max_output_tokens": 8192, "input_cost_per_token": 0, "output_cost_per_token": 0, + "input_cost_per_character": 0, + "output_cost_per_character": 0, "litellm_provider": "vertex_ai-language-models", "mode": "chat", "supports_function_calling": false, @@ -2006,6 +2008,8 @@ "max_output_tokens": 8192, "input_cost_per_token": 0, "output_cost_per_token": 0, + "input_cost_per_character": 0, + "output_cost_per_character": 0, "litellm_provider": "vertex_ai-language-models", "mode": "chat", "supports_function_calling": false,