From 1bd41ab6368f3360d9dcd9314692da21c86f6bdd Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Fri, 20 Oct 2023 14:39:28 -0700 Subject: [PATCH] (feat) add perplexity llms --- litellm/__init__.py | 5 ++++ model_prices_and_context_window.json | 35 ++++++++++++++++++++++++++++ 2 files changed, 40 insertions(+) diff --git a/litellm/__init__.py b/litellm/__init__.py index e69063f62..036d0872b 100644 --- a/litellm/__init__.py +++ b/litellm/__init__.py @@ -100,6 +100,7 @@ nlp_cloud_models: List = [] aleph_alpha_models: List = [] bedrock_models: List = [] deepinfra_models: List = [] +perplexity_models: List = [] for key, value in model_cost.items(): if value.get('litellm_provider') == 'openai': open_ai_chat_completion_models.append(key) @@ -130,6 +131,8 @@ for key, value in model_cost.items(): bedrock_models.append(key) elif value.get('litellm_provider') == 'deepinfra': deepinfra_models.append(key) + elif value.get('litellm_provider') == 'perplexity': + perplexity_models.append(key) # known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary openai_compatible_endpoints: List = [ @@ -234,6 +237,7 @@ model_list = ( + ollama_models + bedrock_models + deepinfra_models + + perplexity_models ) provider_list: List = [ @@ -277,6 +281,7 @@ models_by_provider: dict = { "petals": petals_models, "ollama": ollama_models, "deepinfra": deepinfra_models, + "perplexity": perplexity_models, } # mapping for those models which have larger equivalents diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index e3e9ee810..d8450a715 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -612,5 +612,40 @@ "output_cost_per_token": 0.00000095, "litellm_provider": "deepinfra", "mode": "chat" + }, + "codellama-34b-instruct": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "llama-2-13b-chat": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "llama-2-70b-chat": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "mistral-7b-instruct": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "replit-code-v1.5-3b": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" } }