Update pricing and add cohere refresh models

This commit is contained in:
Jay Alammar 2024-09-06 17:35:59 -04:00
parent 4626c5a365
commit 6d9f6a1b76

View file

@ -890,7 +890,7 @@
"output_cost_per_token": 0.00000061, "output_cost_per_token": 0.00000061,
"litellm_provider": "azure_ai", "litellm_provider": "azure_ai",
"mode": "chat", "mode": "chat",
"source":"https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-8b-instruct-offer?tab=PlansAndPrice" "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-8b-instruct-offer?tab=PlansAndPrice"
}, },
"azure_ai/Meta-Llama-31-70B-Instruct": { "azure_ai/Meta-Llama-31-70B-Instruct": {
"max_tokens": 128000, "max_tokens": 128000,
@ -900,7 +900,7 @@
"output_cost_per_token": 0.00000354, "output_cost_per_token": 0.00000354,
"litellm_provider": "azure_ai", "litellm_provider": "azure_ai",
"mode": "chat", "mode": "chat",
"source":"https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice" "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice"
}, },
"azure_ai/Meta-Llama-31-405B-Instruct": { "azure_ai/Meta-Llama-31-405B-Instruct": {
"max_tokens": 128000, "max_tokens": 128000,
@ -910,7 +910,7 @@
"output_cost_per_token": 0.000016, "output_cost_per_token": 0.000016,
"litellm_provider": "azure_ai", "litellm_provider": "azure_ai",
"mode": "chat", "mode": "chat",
"source":"https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice" "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice"
}, },
"babbage-002": { "babbage-002": {
"max_tokens": 16384, "max_tokens": 16384,
@ -947,7 +947,6 @@
"output_cost_per_token": 0.000002, "output_cost_per_token": 0.000002,
"litellm_provider": "text-completion-openai", "litellm_provider": "text-completion-openai",
"mode": "completion" "mode": "completion"
}, },
"claude-instant-1": { "claude-instant-1": {
"max_tokens": 8191, "max_tokens": 8191,
@ -2432,7 +2431,7 @@
"mode": "embedding", "mode": "embedding",
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing" "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
}, },
"text-multilingual-embedding-preview-0409":{ "text-multilingual-embedding-preview-0409": {
"max_tokens": 3072, "max_tokens": 3072,
"max_input_tokens": 3072, "max_input_tokens": 3072,
"output_vector_size": 768, "output_vector_size": 768,
@ -2665,8 +2664,18 @@
"max_tokens": 4096, "max_tokens": 4096,
"max_input_tokens": 128000, "max_input_tokens": 128000,
"max_output_tokens": 4096, "max_output_tokens": 4096,
"input_cost_per_token": 0.00000050, "input_cost_per_token": 0.00000015,
"output_cost_per_token": 0.0000015, "output_cost_per_token": 0.0000006,
"litellm_provider": "cohere_chat",
"mode": "chat",
"supports_function_calling": true
},
"command-r-08-2024": {
"max_tokens": 4096,
"max_input_tokens": 128000,
"max_output_tokens": 4096,
"input_cost_per_token": 0.00000015,
"output_cost_per_token": 0.0000006,
"litellm_provider": "cohere_chat", "litellm_provider": "cohere_chat",
"mode": "chat", "mode": "chat",
"supports_function_calling": true "supports_function_calling": true
@ -2684,8 +2693,18 @@
"max_tokens": 4096, "max_tokens": 4096,
"max_input_tokens": 128000, "max_input_tokens": 128000,
"max_output_tokens": 4096, "max_output_tokens": 4096,
"input_cost_per_token": 0.000003, "input_cost_per_token": 0.0000025,
"output_cost_per_token": 0.000015, "output_cost_per_token": 0.00001,
"litellm_provider": "cohere_chat",
"mode": "chat",
"supports_function_calling": true
},
"command-r-plus-08-2024": {
"max_tokens": 4096,
"max_input_tokens": 128000,
"max_output_tokens": 4096,
"input_cost_per_token": 0.0000025,
"output_cost_per_token": 0.00001,
"litellm_provider": "cohere_chat", "litellm_provider": "cohere_chat",
"mode": "chat", "mode": "chat",
"supports_function_calling": true "supports_function_calling": true
@ -2707,24 +2726,6 @@
"output_cost_per_token": 0.000015, "output_cost_per_token": 0.000015,
"litellm_provider": "cohere", "litellm_provider": "cohere",
"mode": "completion" "mode": "completion"
},
"command-medium-beta": {
"max_tokens": 4096,
"max_input_tokens": 4096,
"max_output_tokens": 4096,
"input_cost_per_token": 0.000015,
"output_cost_per_token": 0.000015,
"litellm_provider": "cohere",
"mode": "completion"
},
"command-xlarge-beta": {
"max_tokens": 4096,
"max_input_tokens": 4096,
"max_output_tokens": 4096,
"input_cost_per_token": 0.000015,
"output_cost_per_token": 0.000015,
"litellm_provider": "cohere",
"mode": "completion"
}, },
"embed-english-v3.0": { "embed-english-v3.0": {
"max_tokens": 512, "max_tokens": 512,
@ -5331,7 +5332,7 @@
"output_cost_per_token": 0.000001, "output_cost_per_token": 0.000001,
"litellm_provider": "anyscale", "litellm_provider": "anyscale",
"mode": "chat", "mode": "chat",
"source" : "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/codellama-CodeLlama-70b-Instruct-hf" "source": "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/codellama-CodeLlama-70b-Instruct-hf"
}, },
"anyscale/meta-llama/Meta-Llama-3-8B-Instruct": { "anyscale/meta-llama/Meta-Llama-3-8B-Instruct": {
"max_tokens": 8192, "max_tokens": 8192,
@ -5351,7 +5352,7 @@
"output_cost_per_token": 0.00000100, "output_cost_per_token": 0.00000100,
"litellm_provider": "anyscale", "litellm_provider": "anyscale",
"mode": "chat", "mode": "chat",
"source" : "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/meta-llama-Meta-Llama-3-70B-Instruct" "source": "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/meta-llama-Meta-Llama-3-70B-Instruct"
}, },
"cloudflare/@cf/meta/llama-2-7b-chat-fp16": { "cloudflare/@cf/meta/llama-2-7b-chat-fp16": {
"max_tokens": 3072, "max_tokens": 3072,
@ -5494,7 +5495,6 @@
"litellm_provider": "databricks", "litellm_provider": "databricks",
"mode": "chat", "mode": "chat",
"source": "https://www.databricks.com/product/pricing/foundation-model-serving" "source": "https://www.databricks.com/product/pricing/foundation-model-serving"
}, },
"databricks/databricks-mixtral-8x7b-instruct": { "databricks/databricks-mixtral-8x7b-instruct": {
"max_tokens": 4096, "max_tokens": 4096,