mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
Litellm dev 01 29 2025 p1 (#8097)
* Added deprecation date to model_prices_and_context_window.json (#8061) * build(model_prices_and_context_window.json): add xai model pricing --------- Co-authored-by: Юрій <yurikizio@gmail.com>
This commit is contained in:
parent
d10bd67d33
commit
a57fad1e29
2 changed files with 264 additions and 54 deletions
|
@ -14,7 +14,8 @@
|
|||
"supports_audio_output": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD"
|
||||
},
|
||||
"omni-moderation-latest": {
|
||||
"max_tokens": 32768,
|
||||
|
@ -441,7 +442,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "2025-06-06"
|
||||
},
|
||||
"gpt-4-32k": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -540,7 +542,8 @@
|
|||
"mode": "chat",
|
||||
"supports_vision": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "2024-12-06"
|
||||
},
|
||||
"gpt-4-1106-vision-preview": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -552,7 +555,8 @@
|
|||
"mode": "chat",
|
||||
"supports_vision": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "2024-12-06"
|
||||
},
|
||||
"gpt-3.5-turbo": {
|
||||
"max_tokens": 4097,
|
||||
|
@ -1223,7 +1227,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-03-31"
|
||||
},
|
||||
"azure/gpt-35-turbo-0613": {
|
||||
"max_tokens": 4097,
|
||||
|
@ -1234,7 +1239,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-02-13"
|
||||
},
|
||||
"azure/gpt-35-turbo-0301": {
|
||||
"max_tokens": 4097,
|
||||
|
@ -1245,7 +1251,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-02-13"
|
||||
},
|
||||
"azure/gpt-35-turbo-0125": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1256,7 +1263,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-03-31"
|
||||
},
|
||||
"azure/gpt-35-turbo-16k": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -2043,6 +2051,84 @@
|
|||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-vision-1212": {
|
||||
"max_tokens": 32768,
|
||||
"max_input_tokens": 32768,
|
||||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"input_cost_per_image": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-vision-latest": {
|
||||
"max_tokens": 32768,
|
||||
"max_input_tokens": 32768,
|
||||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"input_cost_per_image": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-vision": {
|
||||
"max_tokens": 32768,
|
||||
"max_input_tokens": 32768,
|
||||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"input_cost_per_image": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-vision-beta": {
|
||||
"max_tokens": 8192,
|
||||
"max_input_tokens": 8192,
|
||||
"max_output_tokens": 8192,
|
||||
"input_cost_per_token": 0.000005,
|
||||
"input_cost_per_image": 0.000005,
|
||||
"output_cost_per_token": 0.000015,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-1212": {
|
||||
"max_tokens": 131072,
|
||||
"max_input_tokens": 131072,
|
||||
"max_output_tokens": 131072,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
},
|
||||
"xai/grok-2": {
|
||||
"max_tokens": 131072,
|
||||
"max_input_tokens": 131072,
|
||||
"max_output_tokens": 131072,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
},
|
||||
"xai/grok-2-latest": {
|
||||
"max_tokens": 131072,
|
||||
"max_input_tokens": 131072,
|
||||
"max_output_tokens": 131072,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
},
|
||||
"deepseek/deepseek-coder": {
|
||||
"max_tokens": 4096,
|
||||
"max_input_tokens": 128000,
|
||||
|
@ -2354,7 +2440,8 @@
|
|||
"tool_use_system_prompt_tokens": 264,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-03-01"
|
||||
},
|
||||
"claude-3-5-haiku-20241022": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2370,7 +2457,8 @@
|
|||
"tool_use_system_prompt_tokens": 264,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-10-01"
|
||||
},
|
||||
"claude-3-opus-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -2387,7 +2475,8 @@
|
|||
"tool_use_system_prompt_tokens": 395,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-03-01"
|
||||
},
|
||||
"claude-3-sonnet-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -2402,7 +2491,8 @@
|
|||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-07-21"
|
||||
},
|
||||
"claude-3-5-sonnet-20240620": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2419,7 +2509,8 @@
|
|||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-06-01"
|
||||
},
|
||||
"claude-3-5-sonnet-20241022": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2437,7 +2528,8 @@
|
|||
"supports_assistant_prefill": true,
|
||||
"supports_pdf_input": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-10-01"
|
||||
},
|
||||
"text-bison": {
|
||||
"max_tokens": 2048,
|
||||
|
@ -2547,7 +2639,8 @@
|
|||
"output_cost_per_character": 0.0000005,
|
||||
"litellm_provider": "vertex_ai-chat-models",
|
||||
"mode": "chat",
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"chat-bison-32k": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2788,7 +2881,8 @@
|
|||
"litellm_provider": "vertex_ai-language-models",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"gemini-1.0-ultra": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2833,7 +2927,8 @@
|
|||
"litellm_provider": "vertex_ai-language-models",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"gemini-1.5-pro": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2915,7 +3010,8 @@
|
|||
"supports_function_calling": true,
|
||||
"supports_tool_choice": true,
|
||||
"supports_response_schema": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini-1.5-pro-preview-0514": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -3120,7 +3216,8 @@
|
|||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"supports_response_schema": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini-1.5-flash-preview-0514": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -3224,7 +3321,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"medlm-medium": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -3930,7 +4028,8 @@
|
|||
"supports_prompt_caching": true,
|
||||
"tpm": 4000000,
|
||||
"rpm": 2000,
|
||||
"source": "https://ai.google.dev/pricing"
|
||||
"source": "https://ai.google.dev/pricing",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini/gemini-1.5-flash": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -4206,7 +4305,8 @@
|
|||
"supports_prompt_caching": true,
|
||||
"tpm": 4000000,
|
||||
"rpm": 1000,
|
||||
"source": "https://ai.google.dev/pricing"
|
||||
"source": "https://ai.google.dev/pricing",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini/gemini-1.5-pro-exp-0801": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -7342,7 +7442,8 @@
|
|||
"input_cost_per_token": 0.000005,
|
||||
"output_cost_per_token": 0.000005,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-large-128k-online": {
|
||||
"max_tokens": 127072,
|
||||
|
@ -7351,7 +7452,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000001,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-large-128k-chat": {
|
||||
"max_tokens": 131072,
|
||||
|
@ -7360,7 +7462,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000001,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-small-128k-chat": {
|
||||
"max_tokens": 131072,
|
||||
|
@ -7369,7 +7472,8 @@
|
|||
"input_cost_per_token": 0.0000002,
|
||||
"output_cost_per_token": 0.0000002,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-small-128k-online": {
|
||||
"max_tokens": 127072,
|
||||
|
@ -7378,7 +7482,8 @@
|
|||
"input_cost_per_token": 0.0000002,
|
||||
"output_cost_per_token": 0.0000002,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat" ,
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/pplx-7b-chat": {
|
||||
"max_tokens": 8192,
|
||||
|
|
|
@ -14,7 +14,8 @@
|
|||
"supports_audio_output": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD"
|
||||
},
|
||||
"omni-moderation-latest": {
|
||||
"max_tokens": 32768,
|
||||
|
@ -441,7 +442,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "2025-06-06"
|
||||
},
|
||||
"gpt-4-32k": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -540,7 +542,8 @@
|
|||
"mode": "chat",
|
||||
"supports_vision": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "2024-12-06"
|
||||
},
|
||||
"gpt-4-1106-vision-preview": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -552,7 +555,8 @@
|
|||
"mode": "chat",
|
||||
"supports_vision": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_system_messages": true
|
||||
"supports_system_messages": true,
|
||||
"deprecation_date": "2024-12-06"
|
||||
},
|
||||
"gpt-3.5-turbo": {
|
||||
"max_tokens": 4097,
|
||||
|
@ -1223,7 +1227,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-03-31"
|
||||
},
|
||||
"azure/gpt-35-turbo-0613": {
|
||||
"max_tokens": 4097,
|
||||
|
@ -1234,7 +1239,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-02-13"
|
||||
},
|
||||
"azure/gpt-35-turbo-0301": {
|
||||
"max_tokens": 4097,
|
||||
|
@ -1245,7 +1251,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-02-13"
|
||||
},
|
||||
"azure/gpt-35-turbo-0125": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1256,7 +1263,8 @@
|
|||
"litellm_provider": "azure",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true
|
||||
"supports_parallel_function_calling": true,
|
||||
"deprecation_date": "2025-03-31"
|
||||
},
|
||||
"azure/gpt-35-turbo-16k": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -2043,6 +2051,84 @@
|
|||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-vision-1212": {
|
||||
"max_tokens": 32768,
|
||||
"max_input_tokens": 32768,
|
||||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"input_cost_per_image": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-vision-latest": {
|
||||
"max_tokens": 32768,
|
||||
"max_input_tokens": 32768,
|
||||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"input_cost_per_image": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-vision": {
|
||||
"max_tokens": 32768,
|
||||
"max_input_tokens": 32768,
|
||||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"input_cost_per_image": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-vision-beta": {
|
||||
"max_tokens": 8192,
|
||||
"max_input_tokens": 8192,
|
||||
"max_output_tokens": 8192,
|
||||
"input_cost_per_token": 0.000005,
|
||||
"input_cost_per_image": 0.000005,
|
||||
"output_cost_per_token": 0.000015,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true
|
||||
},
|
||||
"xai/grok-2-1212": {
|
||||
"max_tokens": 131072,
|
||||
"max_input_tokens": 131072,
|
||||
"max_output_tokens": 131072,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
},
|
||||
"xai/grok-2": {
|
||||
"max_tokens": 131072,
|
||||
"max_input_tokens": 131072,
|
||||
"max_output_tokens": 131072,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
},
|
||||
"xai/grok-2-latest": {
|
||||
"max_tokens": 131072,
|
||||
"max_input_tokens": 131072,
|
||||
"max_output_tokens": 131072,
|
||||
"input_cost_per_token": 0.000002,
|
||||
"output_cost_per_token": 0.00001,
|
||||
"litellm_provider": "xai",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
},
|
||||
"deepseek/deepseek-coder": {
|
||||
"max_tokens": 4096,
|
||||
"max_input_tokens": 128000,
|
||||
|
@ -2354,7 +2440,8 @@
|
|||
"tool_use_system_prompt_tokens": 264,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-03-01"
|
||||
},
|
||||
"claude-3-5-haiku-20241022": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2370,7 +2457,8 @@
|
|||
"tool_use_system_prompt_tokens": 264,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-10-01"
|
||||
},
|
||||
"claude-3-opus-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -2387,7 +2475,8 @@
|
|||
"tool_use_system_prompt_tokens": 395,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-03-01"
|
||||
},
|
||||
"claude-3-sonnet-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -2402,7 +2491,8 @@
|
|||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-07-21"
|
||||
},
|
||||
"claude-3-5-sonnet-20240620": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2419,7 +2509,8 @@
|
|||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-06-01"
|
||||
},
|
||||
"claude-3-5-sonnet-20241022": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2437,7 +2528,8 @@
|
|||
"supports_assistant_prefill": true,
|
||||
"supports_pdf_input": true,
|
||||
"supports_prompt_caching": true,
|
||||
"supports_response_schema": true
|
||||
"supports_response_schema": true,
|
||||
"deprecation_date": "2025-10-01"
|
||||
},
|
||||
"text-bison": {
|
||||
"max_tokens": 2048,
|
||||
|
@ -2547,7 +2639,8 @@
|
|||
"output_cost_per_character": 0.0000005,
|
||||
"litellm_provider": "vertex_ai-chat-models",
|
||||
"mode": "chat",
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"chat-bison-32k": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2788,7 +2881,8 @@
|
|||
"litellm_provider": "vertex_ai-language-models",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"gemini-1.0-ultra": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2833,7 +2927,8 @@
|
|||
"litellm_provider": "vertex_ai-language-models",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"gemini-1.5-pro": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -2915,7 +3010,8 @@
|
|||
"supports_function_calling": true,
|
||||
"supports_tool_choice": true,
|
||||
"supports_response_schema": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini-1.5-pro-preview-0514": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -3120,7 +3216,8 @@
|
|||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"supports_response_schema": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini-1.5-flash-preview-0514": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -3224,7 +3321,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||
"deprecation_date": "2025-04-09"
|
||||
},
|
||||
"medlm-medium": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -3930,7 +4028,8 @@
|
|||
"supports_prompt_caching": true,
|
||||
"tpm": 4000000,
|
||||
"rpm": 2000,
|
||||
"source": "https://ai.google.dev/pricing"
|
||||
"source": "https://ai.google.dev/pricing",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini/gemini-1.5-flash": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -4206,7 +4305,8 @@
|
|||
"supports_prompt_caching": true,
|
||||
"tpm": 4000000,
|
||||
"rpm": 1000,
|
||||
"source": "https://ai.google.dev/pricing"
|
||||
"source": "https://ai.google.dev/pricing",
|
||||
"deprecation_date": "2025-05-24"
|
||||
},
|
||||
"gemini/gemini-1.5-pro-exp-0801": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -7342,7 +7442,8 @@
|
|||
"input_cost_per_token": 0.000005,
|
||||
"output_cost_per_token": 0.000005,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-large-128k-online": {
|
||||
"max_tokens": 127072,
|
||||
|
@ -7351,7 +7452,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000001,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-large-128k-chat": {
|
||||
"max_tokens": 131072,
|
||||
|
@ -7360,7 +7462,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000001,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-small-128k-chat": {
|
||||
"max_tokens": 131072,
|
||||
|
@ -7369,7 +7472,8 @@
|
|||
"input_cost_per_token": 0.0000002,
|
||||
"output_cost_per_token": 0.0000002,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/llama-3.1-sonar-small-128k-online": {
|
||||
"max_tokens": 127072,
|
||||
|
@ -7378,7 +7482,8 @@
|
|||
"input_cost_per_token": 0.0000002,
|
||||
"output_cost_per_token": 0.0000002,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
"mode": "chat" ,
|
||||
"deprecation_date": "2025-02-22"
|
||||
},
|
||||
"perplexity/pplx-7b-chat": {
|
||||
"max_tokens": 8192,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue