diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 5838935a2..b6ded001c 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -1769,71 +1769,71 @@ "litellm_provider": "deepinfra", "mode": "chat" }, - "perplexity/pplx-7b-chat": { + "perplexity/codellama-34b-instruct": { + "max_tokens": 16384, + "input_cost_per_token": 0.00000035, + "output_cost_per_token": 0.00000140, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "perplexity/codellama-70b-instruct": { + "max_tokens": 16384, + "input_cost_per_token": 0.00000070, + "output_cost_per_token": 0.00000280, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "perplexity/pplx-7b-chat": { "max_tokens": 8192, "input_cost_per_token": 0.00000007, "output_cost_per_token": 0.00000028, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/pplx-34b-chat": { - "max_tokens": 8192, - "input_cost_per_token": 0.00000035, - "output_cost_per_token": 0.00000140, - "litellm_provider": "perplexity", - "mode": "chat" - }, - "perplexity/pplx-70b-chat": { + }, + "perplexity/pplx-70b-chat": { "max_tokens": 4096, "input_cost_per_token": 0.00000070, "output_cost_per_token": 0.00000280, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/pplx-7b-online": { + }, + "perplexity/pplx-7b-online": { "max_tokens": 4096, "input_cost_per_token": 0.0000000, "output_cost_per_token": 0.00000028, "input_cost_per_request": 0.005, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/pplx-70b-online": { + }, + "perplexity/pplx-70b-online": { "max_tokens": 4096, "input_cost_per_token": 0.0000000, "output_cost_per_token": 0.00000280, "input_cost_per_request": 0.005, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/llama-2-13b-chat": { + }, + "perplexity/llama-2-70b-chat": { "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, + "input_cost_per_token": 0.00000070, + "output_cost_per_token": 0.00000280, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/llama-2-70b-chat": { + }, + "perplexity/mistral-7b-instruct": { "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, + "input_cost_per_token": 0.00000007, + "output_cost_per_token": 0.00000028, "litellm_provider": "perplexity", "mode": "chat" - }, - "perplexity/mistral-7b-instruct": { - "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, - "litellm_provider": "perplexity", - "mode": "chat" - }, - "perplexity/replit-code-v1.5-3b": { - "max_tokens": 4096, - "input_cost_per_token": 0.0000000, - "output_cost_per_token": 0.000000, - "litellm_provider": "perplexity", - "mode": "chat" - }, + }, + "perplexity/mixtral-8x7b-instruct": { + "max_tokens": 4096, + "input_cost_per_token": 0.00000007, + "output_cost_per_token": 0.00000028, + "litellm_provider": "perplexity", + "mode": "chat" + }, "anyscale/mistralai/Mistral-7B-Instruct-v0.1": { "max_tokens": 16384, "input_cost_per_token": 0.00000015,