diff --git a/litellm/__init__.py b/litellm/__init__.py index e69063f62d..036d0872b9 100644 --- a/litellm/__init__.py +++ b/litellm/__init__.py @@ -100,6 +100,7 @@ nlp_cloud_models: List = [] aleph_alpha_models: List = [] bedrock_models: List = [] deepinfra_models: List = [] +perplexity_models: List = [] for key, value in model_cost.items(): if value.get('litellm_provider') == 'openai': open_ai_chat_completion_models.append(key) @@ -130,6 +131,8 @@ for key, value in model_cost.items(): bedrock_models.append(key) elif value.get('litellm_provider') == 'deepinfra': deepinfra_models.append(key) + elif value.get('litellm_provider') == 'perplexity': + perplexity_models.append(key) # known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary openai_compatible_endpoints: List = [ @@ -234,6 +237,7 @@ model_list = ( + ollama_models + bedrock_models + deepinfra_models + + perplexity_models ) provider_list: List = [ @@ -277,6 +281,7 @@ models_by_provider: dict = { "petals": petals_models, "ollama": ollama_models, "deepinfra": deepinfra_models, + "perplexity": perplexity_models, } # mapping for those models which have larger equivalents diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index e3e9ee8109..d8450a7152 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -612,5 +612,40 @@ "output_cost_per_token": 0.00000095, "litellm_provider": "deepinfra", "mode": "chat" + }, + "codellama-34b-instruct": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "llama-2-13b-chat": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "llama-2-70b-chat": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "mistral-7b-instruct": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" + }, + "replit-code-v1.5-3b": { + "max_tokens": 4096, + "input_cost_per_token": 0.0000000, + "output_cost_per_token": 0.000000, + "litellm_provider": "perplexity", + "mode": "chat" } }