diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 3139f050f..c6e99f805 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -1045,6 +1045,7 @@ "max_input_tokens": 32000, "max_output_tokens": 4096, "input_cost_per_token": 0.00000014, + "input_cost_per_token_cache_hit": 0.000000014, "output_cost_per_token": 0.00000028, "litellm_provider": "deepseek", "mode": "chat" @@ -1094,6 +1095,7 @@ "max_input_tokens": 128000, "max_output_tokens": 4096, "input_cost_per_token": 0.00000014, + "input_cost_per_token_cache_hit": 0.000000014, "output_cost_per_token": 0.00000028, "litellm_provider": "deepseek", "mode": "chat" diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 3139f050f..c6e99f805 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -1045,6 +1045,7 @@ "max_input_tokens": 32000, "max_output_tokens": 4096, "input_cost_per_token": 0.00000014, + "input_cost_per_token_cache_hit": 0.000000014, "output_cost_per_token": 0.00000028, "litellm_provider": "deepseek", "mode": "chat" @@ -1094,6 +1095,7 @@ "max_input_tokens": 128000, "max_output_tokens": 4096, "input_cost_per_token": 0.00000014, + "input_cost_per_token_cache_hit": 0.000000014, "output_cost_per_token": 0.00000028, "litellm_provider": "deepseek", "mode": "chat"