Litellm dev 01 29 2025 p1 (#8097)

* Added deprecation date to model_prices_and_context_window.json (#8061)

* build(model_prices_and_context_window.json): add xai model pricing

---------

Co-authored-by: Юрій <yurikizio@gmail.com>
This commit is contained in:
Krish Dholakia 2025-01-29 19:50:40 -08:00 committed by GitHub
parent d10bd67d33
commit a57fad1e29
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2 changed files with 264 additions and 54 deletions

View file

@ -14,7 +14,8 @@
"supports_audio_output": true,
"supports_prompt_caching": true,
"supports_response_schema": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD"
},
"omni-moderation-latest": {
"max_tokens": 32768,
@ -441,7 +442,8 @@
"mode": "chat",
"supports_function_calling": true,
"supports_prompt_caching": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "2025-06-06"
},
"gpt-4-32k": {
"max_tokens": 4096,
@ -540,7 +542,8 @@
"mode": "chat",
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "2024-12-06"
},
"gpt-4-1106-vision-preview": {
"max_tokens": 4096,
@ -552,7 +555,8 @@
"mode": "chat",
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "2024-12-06"
},
"gpt-3.5-turbo": {
"max_tokens": 4097,
@ -1223,7 +1227,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-03-31"
},
"azure/gpt-35-turbo-0613": {
"max_tokens": 4097,
@ -1234,7 +1239,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-02-13"
},
"azure/gpt-35-turbo-0301": {
"max_tokens": 4097,
@ -1245,7 +1251,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-02-13"
},
"azure/gpt-35-turbo-0125": {
"max_tokens": 4096,
@ -1256,7 +1263,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-03-31"
},
"azure/gpt-35-turbo-16k": {
"max_tokens": 4096,
@ -2043,6 +2051,84 @@
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-vision-1212": {
"max_tokens": 32768,
"max_input_tokens": 32768,
"max_output_tokens": 32768,
"input_cost_per_token": 0.000002,
"input_cost_per_image": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-vision-latest": {
"max_tokens": 32768,
"max_input_tokens": 32768,
"max_output_tokens": 32768,
"input_cost_per_token": 0.000002,
"input_cost_per_image": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-vision": {
"max_tokens": 32768,
"max_input_tokens": 32768,
"max_output_tokens": 32768,
"input_cost_per_token": 0.000002,
"input_cost_per_image": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-vision-beta": {
"max_tokens": 8192,
"max_input_tokens": 8192,
"max_output_tokens": 8192,
"input_cost_per_token": 0.000005,
"input_cost_per_image": 0.000005,
"output_cost_per_token": 0.000015,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-1212": {
"max_tokens": 131072,
"max_input_tokens": 131072,
"max_output_tokens": 131072,
"input_cost_per_token": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true
},
"xai/grok-2": {
"max_tokens": 131072,
"max_input_tokens": 131072,
"max_output_tokens": 131072,
"input_cost_per_token": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true
},
"xai/grok-2-latest": {
"max_tokens": 131072,
"max_input_tokens": 131072,
"max_output_tokens": 131072,
"input_cost_per_token": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true
},
"deepseek/deepseek-coder": {
"max_tokens": 4096,
"max_input_tokens": 128000,
@ -2354,7 +2440,8 @@
"tool_use_system_prompt_tokens": 264,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-03-01"
},
"claude-3-5-haiku-20241022": {
"max_tokens": 8192,
@ -2370,7 +2457,8 @@
"tool_use_system_prompt_tokens": 264,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-10-01"
},
"claude-3-opus-20240229": {
"max_tokens": 4096,
@ -2387,7 +2475,8 @@
"tool_use_system_prompt_tokens": 395,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-03-01"
},
"claude-3-sonnet-20240229": {
"max_tokens": 4096,
@ -2402,7 +2491,8 @@
"tool_use_system_prompt_tokens": 159,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-07-21"
},
"claude-3-5-sonnet-20240620": {
"max_tokens": 8192,
@ -2419,7 +2509,8 @@
"tool_use_system_prompt_tokens": 159,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-06-01"
},
"claude-3-5-sonnet-20241022": {
"max_tokens": 8192,
@ -2437,7 +2528,8 @@
"supports_assistant_prefill": true,
"supports_pdf_input": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-10-01"
},
"text-bison": {
"max_tokens": 2048,
@ -2547,7 +2639,8 @@
"output_cost_per_character": 0.0000005,
"litellm_provider": "vertex_ai-chat-models",
"mode": "chat",
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"chat-bison-32k": {
"max_tokens": 8192,
@ -2788,7 +2881,8 @@
"litellm_provider": "vertex_ai-language-models",
"mode": "chat",
"supports_function_calling": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"gemini-1.0-ultra": {
"max_tokens": 8192,
@ -2833,7 +2927,8 @@
"litellm_provider": "vertex_ai-language-models",
"mode": "chat",
"supports_function_calling": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"gemini-1.5-pro": {
"max_tokens": 8192,
@ -2915,7 +3010,8 @@
"supports_function_calling": true,
"supports_tool_choice": true,
"supports_response_schema": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-05-24"
},
"gemini-1.5-pro-preview-0514": {
"max_tokens": 8192,
@ -3120,7 +3216,8 @@
"supports_function_calling": true,
"supports_vision": true,
"supports_response_schema": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-05-24"
},
"gemini-1.5-flash-preview-0514": {
"max_tokens": 8192,
@ -3224,7 +3321,8 @@
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"medlm-medium": {
"max_tokens": 8192,
@ -3930,7 +4028,8 @@
"supports_prompt_caching": true,
"tpm": 4000000,
"rpm": 2000,
"source": "https://ai.google.dev/pricing"
"source": "https://ai.google.dev/pricing",
"deprecation_date": "2025-05-24"
},
"gemini/gemini-1.5-flash": {
"max_tokens": 8192,
@ -4206,7 +4305,8 @@
"supports_prompt_caching": true,
"tpm": 4000000,
"rpm": 1000,
"source": "https://ai.google.dev/pricing"
"source": "https://ai.google.dev/pricing",
"deprecation_date": "2025-05-24"
},
"gemini/gemini-1.5-pro-exp-0801": {
"max_tokens": 8192,
@ -7342,7 +7442,8 @@
"input_cost_per_token": 0.000005,
"output_cost_per_token": 0.000005,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-large-128k-online": {
"max_tokens": 127072,
@ -7351,7 +7452,8 @@
"input_cost_per_token": 0.000001,
"output_cost_per_token": 0.000001,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-large-128k-chat": {
"max_tokens": 131072,
@ -7360,7 +7462,8 @@
"input_cost_per_token": 0.000001,
"output_cost_per_token": 0.000001,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-small-128k-chat": {
"max_tokens": 131072,
@ -7369,7 +7472,8 @@
"input_cost_per_token": 0.0000002,
"output_cost_per_token": 0.0000002,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-small-128k-online": {
"max_tokens": 127072,
@ -7378,7 +7482,8 @@
"input_cost_per_token": 0.0000002,
"output_cost_per_token": 0.0000002,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat" ,
"deprecation_date": "2025-02-22"
},
"perplexity/pplx-7b-chat": {
"max_tokens": 8192,

View file

@ -14,7 +14,8 @@
"supports_audio_output": true,
"supports_prompt_caching": true,
"supports_response_schema": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD"
},
"omni-moderation-latest": {
"max_tokens": 32768,
@ -441,7 +442,8 @@
"mode": "chat",
"supports_function_calling": true,
"supports_prompt_caching": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "2025-06-06"
},
"gpt-4-32k": {
"max_tokens": 4096,
@ -540,7 +542,8 @@
"mode": "chat",
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "2024-12-06"
},
"gpt-4-1106-vision-preview": {
"max_tokens": 4096,
@ -552,7 +555,8 @@
"mode": "chat",
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true
"supports_system_messages": true,
"deprecation_date": "2024-12-06"
},
"gpt-3.5-turbo": {
"max_tokens": 4097,
@ -1223,7 +1227,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-03-31"
},
"azure/gpt-35-turbo-0613": {
"max_tokens": 4097,
@ -1234,7 +1239,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-02-13"
},
"azure/gpt-35-turbo-0301": {
"max_tokens": 4097,
@ -1245,7 +1251,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-02-13"
},
"azure/gpt-35-turbo-0125": {
"max_tokens": 4096,
@ -1256,7 +1263,8 @@
"litellm_provider": "azure",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true
"supports_parallel_function_calling": true,
"deprecation_date": "2025-03-31"
},
"azure/gpt-35-turbo-16k": {
"max_tokens": 4096,
@ -2043,6 +2051,84 @@
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-vision-1212": {
"max_tokens": 32768,
"max_input_tokens": 32768,
"max_output_tokens": 32768,
"input_cost_per_token": 0.000002,
"input_cost_per_image": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-vision-latest": {
"max_tokens": 32768,
"max_input_tokens": 32768,
"max_output_tokens": 32768,
"input_cost_per_token": 0.000002,
"input_cost_per_image": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-vision": {
"max_tokens": 32768,
"max_input_tokens": 32768,
"max_output_tokens": 32768,
"input_cost_per_token": 0.000002,
"input_cost_per_image": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-vision-beta": {
"max_tokens": 8192,
"max_input_tokens": 8192,
"max_output_tokens": 8192,
"input_cost_per_token": 0.000005,
"input_cost_per_image": 0.000005,
"output_cost_per_token": 0.000015,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true
},
"xai/grok-2-1212": {
"max_tokens": 131072,
"max_input_tokens": 131072,
"max_output_tokens": 131072,
"input_cost_per_token": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true
},
"xai/grok-2": {
"max_tokens": 131072,
"max_input_tokens": 131072,
"max_output_tokens": 131072,
"input_cost_per_token": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true
},
"xai/grok-2-latest": {
"max_tokens": 131072,
"max_input_tokens": 131072,
"max_output_tokens": 131072,
"input_cost_per_token": 0.000002,
"output_cost_per_token": 0.00001,
"litellm_provider": "xai",
"mode": "chat",
"supports_function_calling": true
},
"deepseek/deepseek-coder": {
"max_tokens": 4096,
"max_input_tokens": 128000,
@ -2354,7 +2440,8 @@
"tool_use_system_prompt_tokens": 264,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-03-01"
},
"claude-3-5-haiku-20241022": {
"max_tokens": 8192,
@ -2370,7 +2457,8 @@
"tool_use_system_prompt_tokens": 264,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-10-01"
},
"claude-3-opus-20240229": {
"max_tokens": 4096,
@ -2387,7 +2475,8 @@
"tool_use_system_prompt_tokens": 395,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-03-01"
},
"claude-3-sonnet-20240229": {
"max_tokens": 4096,
@ -2402,7 +2491,8 @@
"tool_use_system_prompt_tokens": 159,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-07-21"
},
"claude-3-5-sonnet-20240620": {
"max_tokens": 8192,
@ -2419,7 +2509,8 @@
"tool_use_system_prompt_tokens": 159,
"supports_assistant_prefill": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-06-01"
},
"claude-3-5-sonnet-20241022": {
"max_tokens": 8192,
@ -2437,7 +2528,8 @@
"supports_assistant_prefill": true,
"supports_pdf_input": true,
"supports_prompt_caching": true,
"supports_response_schema": true
"supports_response_schema": true,
"deprecation_date": "2025-10-01"
},
"text-bison": {
"max_tokens": 2048,
@ -2547,7 +2639,8 @@
"output_cost_per_character": 0.0000005,
"litellm_provider": "vertex_ai-chat-models",
"mode": "chat",
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"chat-bison-32k": {
"max_tokens": 8192,
@ -2788,7 +2881,8 @@
"litellm_provider": "vertex_ai-language-models",
"mode": "chat",
"supports_function_calling": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"gemini-1.0-ultra": {
"max_tokens": 8192,
@ -2833,7 +2927,8 @@
"litellm_provider": "vertex_ai-language-models",
"mode": "chat",
"supports_function_calling": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"gemini-1.5-pro": {
"max_tokens": 8192,
@ -2915,7 +3010,8 @@
"supports_function_calling": true,
"supports_tool_choice": true,
"supports_response_schema": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-05-24"
},
"gemini-1.5-pro-preview-0514": {
"max_tokens": 8192,
@ -3120,7 +3216,8 @@
"supports_function_calling": true,
"supports_vision": true,
"supports_response_schema": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-05-24"
},
"gemini-1.5-flash-preview-0514": {
"max_tokens": 8192,
@ -3224,7 +3321,8 @@
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true,
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"deprecation_date": "2025-04-09"
},
"medlm-medium": {
"max_tokens": 8192,
@ -3930,7 +4028,8 @@
"supports_prompt_caching": true,
"tpm": 4000000,
"rpm": 2000,
"source": "https://ai.google.dev/pricing"
"source": "https://ai.google.dev/pricing",
"deprecation_date": "2025-05-24"
},
"gemini/gemini-1.5-flash": {
"max_tokens": 8192,
@ -4206,7 +4305,8 @@
"supports_prompt_caching": true,
"tpm": 4000000,
"rpm": 1000,
"source": "https://ai.google.dev/pricing"
"source": "https://ai.google.dev/pricing",
"deprecation_date": "2025-05-24"
},
"gemini/gemini-1.5-pro-exp-0801": {
"max_tokens": 8192,
@ -7342,7 +7442,8 @@
"input_cost_per_token": 0.000005,
"output_cost_per_token": 0.000005,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-large-128k-online": {
"max_tokens": 127072,
@ -7351,7 +7452,8 @@
"input_cost_per_token": 0.000001,
"output_cost_per_token": 0.000001,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-large-128k-chat": {
"max_tokens": 131072,
@ -7360,7 +7462,8 @@
"input_cost_per_token": 0.000001,
"output_cost_per_token": 0.000001,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-small-128k-chat": {
"max_tokens": 131072,
@ -7369,7 +7472,8 @@
"input_cost_per_token": 0.0000002,
"output_cost_per_token": 0.0000002,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat",
"deprecation_date": "2025-02-22"
},
"perplexity/llama-3.1-sonar-small-128k-online": {
"max_tokens": 127072,
@ -7378,7 +7482,8 @@
"input_cost_per_token": 0.0000002,
"output_cost_per_token": 0.0000002,
"litellm_provider": "perplexity",
"mode": "chat"
"mode": "chat" ,
"deprecation_date": "2025-02-22"
},
"perplexity/pplx-7b-chat": {
"max_tokens": 8192,