build(model_prices_and_context_window.json): add gpt-4.1-mini and gpt-4.1-nano model support

This commit is contained in:
Krrish Dholakia 2025-04-14 12:12:40 -07:00
parent b2d38a1323
commit 4daa76aa61
2 changed files with 192 additions and 4 deletions

View file

@ -84,7 +84,8 @@
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-2025-04-14": {
"max_tokens": 32768,
@ -106,7 +107,100 @@
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-mini": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.4e-6,
"output_cost_per_token": 1.6e-6,
"input_cost_per_token_batches": 0.2e-6,
"output_cost_per_token_batches": 0.8e-6,
"cache_read_input_token_cost": 0.1e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-mini-2025-04-14": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.4e-6,
"output_cost_per_token": 1.6e-6,
"input_cost_per_token_batches": 0.2e-6,
"output_cost_per_token_batches": 0.8e-6,
"cache_read_input_token_cost": 0.1e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-nano": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.1e-6,
"output_cost_per_token": 0.4e-6,
"input_cost_per_token_batches": 0.05e-6,
"output_cost_per_token_batches": 0.2e-6,
"cache_read_input_token_cost": 0.025e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-nano-2025-04-14": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.1e-6,
"output_cost_per_token": 0.4e-6,
"input_cost_per_token_batches": 0.05e-6,
"output_cost_per_token_batches": 0.2e-6,
"cache_read_input_token_cost": 0.025e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4o": {
"max_tokens": 16384,

View file

@ -84,7 +84,8 @@
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-2025-04-14": {
"max_tokens": 32768,
@ -106,7 +107,100 @@
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-mini": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.4e-6,
"output_cost_per_token": 1.6e-6,
"input_cost_per_token_batches": 0.2e-6,
"output_cost_per_token_batches": 0.8e-6,
"cache_read_input_token_cost": 0.1e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-mini-2025-04-14": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.4e-6,
"output_cost_per_token": 1.6e-6,
"input_cost_per_token_batches": 0.2e-6,
"output_cost_per_token_batches": 0.8e-6,
"cache_read_input_token_cost": 0.1e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-nano": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.1e-6,
"output_cost_per_token": 0.4e-6,
"input_cost_per_token_batches": 0.05e-6,
"output_cost_per_token_batches": 0.2e-6,
"cache_read_input_token_cost": 0.025e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4.1-nano-2025-04-14": {
"max_tokens": 32768,
"max_input_tokens": 1047576,
"max_output_tokens": 32768,
"input_cost_per_token": 0.1e-6,
"output_cost_per_token": 0.4e-6,
"input_cost_per_token_batches": 0.05e-6,
"output_cost_per_token_batches": 0.2e-6,
"cache_read_input_token_cost": 0.025e-6,
"litellm_provider": "openai",
"mode": "chat",
"supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"],
"supported_modalities": ["text", "image"],
"supported_output_modalities": ["text"],
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true,
"supports_native_streaming": true
},
"gpt-4o": {
"max_tokens": 16384,