diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 6b075e111..adb1f0ac7 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -11,42 +11,6 @@ "supports_parallel_function_calling": true, "supports_vision": true }, - "o1-preview-2024-09-12": { - "max_tokens": 4096, - "max_input_tokens": 128000, - "max_output_tokens": 32768, - "input_cost_per_token": 0.000015, - "output_cost_per_token": 0.00006, - "litellm_provider": "openai", - "mode": "chat" - }, - "o1-preview": { - "max_tokens": 4096, - "max_input_tokens": 128000, - "max_output_tokens": 32768, - "input_cost_per_token": 0.000015, - "output_cost_per_token": 0.00006, - "litellm_provider": "openai", - "mode": "chat" - }, - "o1-mini": { - "max_tokens": 4096, - "max_input_tokens": 128000, - "max_output_tokens": 65536, - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.000012, - "litellm_provider": "openai", - "mode": "chat" - }, - "o1-mini-2024-09-12": { - "max_tokens": 4096, - "max_input_tokens": 128000, - "max_output_tokens": 65536, - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.000012, - "litellm_provider": "openai", - "mode": "chat" - }, "gpt-4": { "max_tokens": 4096, "max_input_tokens": 8192, @@ -94,9 +58,9 @@ "supports_vision": true }, "o1-mini": { - "max_tokens": 16384, + "max_tokens": 65536, "max_input_tokens": 128000, - "max_output_tokens": 16384, + "max_output_tokens": 65536, "input_cost_per_token": 0.000003, "output_cost_per_token": 0.000012, "litellm_provider": "openai", @@ -106,9 +70,9 @@ "supports_vision": true }, "o1-mini-2024-09-12": { - "max_tokens": 16384, + "max_tokens": 65536, "max_input_tokens": 128000, - "max_output_tokens": 16384, + "max_output_tokens": 65536, "input_cost_per_token": 0.000003, "output_cost_per_token": 0.000012, "litellm_provider": "openai", @@ -118,9 +82,9 @@ "supports_vision": true }, "o1-preview": { - "max_tokens": 16384, + "max_tokens": 32768, "max_input_tokens": 128000, - "max_output_tokens": 16384, + "max_output_tokens": 32768, "input_cost_per_token": 0.000015, "output_cost_per_token": 0.000060, "litellm_provider": "openai", @@ -130,9 +94,9 @@ "supports_vision": true }, "o1-preview-2024-09-12": { - "max_tokens": 16384, + "max_tokens": 32768, "max_input_tokens": 128000, - "max_output_tokens": 16384, + "max_output_tokens": 32768, "input_cost_per_token": 0.000015, "output_cost_per_token": 0.000060, "litellm_provider": "openai",