From e832492423beae14755fc932b0e08f50cb6f82b4 Mon Sep 17 00:00:00 2001 From: Toni Engelhardt Date: Mon, 5 Feb 2024 16:46:10 +0000 Subject: [PATCH] Update Perplexity models in model_prices_and_context_window.json According to https://docs.perplexity.ai/docs/model-cards --- model_prices_and_context_window.json | 74 ++++++++++++++-------------- 1 file changed, 37 insertions(+), 37 deletions(-) diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 5838935a2..b6ded001c 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -1769,71 +1769,71 @@ "litellm_provider": "deepinfra", "mode": "chat" }, - "perplexity/pplx-7b-chat": { + "perplexity/codellama-34b-instruct": { + "max_tokens": 16384, + "input_cost_per_token": 0.00000035, + "output_cost_per_token": 0.00000140, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "perplexity/codellama-70b-instruct": { + "max_tokens": 16384, + "input_cost_per_token": 0.00000070, + "output_cost_per_token": 0.00000280, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "perplexity/pplx-7b-chat": { "max_tokens": 8192, "input_cost_per_token": 0.00000007, "output_cost_per_token": 0.00000028, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/pplx-34b-chat": { - "max_tokens": 8192, - "input_cost_per_token": 0.00000035, - "output_cost_per_token": 0.00000140, - "litellm_provider": "perplexity", - "mode": "chat" - }, - "perplexity/pplx-70b-chat": { + }, + "perplexity/pplx-70b-chat": { "max_tokens": 4096, "input_cost_per_token": 0.00000070, "output_cost_per_token": 0.00000280, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/pplx-7b-online": { + }, + "perplexity/pplx-7b-online": { "max_tokens": 4096, "input_cost_per_token": 0.0000000, "output_cost_per_token": 0.00000028, "input_cost_per_request": 0.005, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/pplx-70b-online": { + }, + "perplexity/pplx-70b-online": { "max_tokens": 4096, "input_cost_per_token": 0.0000000, "output_cost_per_token": 0.00000280, "input_cost_per_request": 0.005, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/llama-2-13b-chat": { + }, + "perplexity/llama-2-70b-chat": { "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, + "input_cost_per_token": 0.00000070, + "output_cost_per_token": 0.00000280, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/llama-2-70b-chat": { + }, + "perplexity/mistral-7b-instruct": { "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, + "input_cost_per_token": 0.00000007, + "output_cost_per_token": 0.00000028, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/mistral-7b-instruct": { - "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, - "litellm_provider": "perplexity", - "mode": "chat" - }, - "perplexity/replit-code-v1.5-3b": { - "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, - "litellm_provider": "perplexity", - "mode": "chat" - }, + }, + "perplexity/mixtral-8x7b-instruct": { + "max_tokens": 4096, + "input_cost_per_token": 0.00000007, + "output_cost_per_token": 0.00000028, + "litellm_provider": "perplexity", + "mode": "chat" + }, "anyscale/mistralai/Mistral-7B-Instruct-v0.1": { "max_tokens": 16384, "input_cost_per_token": 0.00000015,