(feat) add perplexity llms

This commit is contained in:
ishaan-jaff 2023-10-20 14:39:28 -07:00
parent 0b94e968e1
commit 1bd41ab636
2 changed files with 40 additions and 0 deletions

View file

@ -100,6 +100,7 @@ nlp_cloud_models: List = []
aleph_alpha_models: List = [] aleph_alpha_models: List = []
bedrock_models: List = [] bedrock_models: List = []
deepinfra_models: List = [] deepinfra_models: List = []
perplexity_models: List = []
for key, value in model_cost.items(): for key, value in model_cost.items():
if value.get('litellm_provider') == 'openai': if value.get('litellm_provider') == 'openai':
open_ai_chat_completion_models.append(key) open_ai_chat_completion_models.append(key)
@ -130,6 +131,8 @@ for key, value in model_cost.items():
bedrock_models.append(key) bedrock_models.append(key)
elif value.get('litellm_provider') == 'deepinfra': elif value.get('litellm_provider') == 'deepinfra':
deepinfra_models.append(key) deepinfra_models.append(key)
elif value.get('litellm_provider') == 'perplexity':
perplexity_models.append(key)
# known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary # known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary
openai_compatible_endpoints: List = [ openai_compatible_endpoints: List = [
@ -234,6 +237,7 @@ model_list = (
+ ollama_models + ollama_models
+ bedrock_models + bedrock_models
+ deepinfra_models + deepinfra_models
+ perplexity_models
) )
provider_list: List = [ provider_list: List = [
@ -277,6 +281,7 @@ models_by_provider: dict = {
"petals": petals_models, "petals": petals_models,
"ollama": ollama_models, "ollama": ollama_models,
"deepinfra": deepinfra_models, "deepinfra": deepinfra_models,
"perplexity": perplexity_models,
} }
# mapping for those models which have larger equivalents # mapping for those models which have larger equivalents

View file

@ -612,5 +612,40 @@
"output_cost_per_token": 0.00000095, "output_cost_per_token": 0.00000095,
"litellm_provider": "deepinfra", "litellm_provider": "deepinfra",
"mode": "chat" "mode": "chat"
},
"codellama-34b-instruct": {
"max_tokens": 4096,
"input_cost_per_token": 0.0000000,
"output_cost_per_token": 0.000000,
"litellm_provider": "perplexity",
"mode": "chat"
},
"llama-2-13b-chat": {
"max_tokens": 4096,
"input_cost_per_token": 0.0000000,
"output_cost_per_token": 0.000000,
"litellm_provider": "perplexity",
"mode": "chat"
},
"llama-2-70b-chat": {
"max_tokens": 4096,
"input_cost_per_token": 0.0000000,
"output_cost_per_token": 0.000000,
"litellm_provider": "perplexity",
"mode": "chat"
},
"mistral-7b-instruct": {
"max_tokens": 4096,
"input_cost_per_token": 0.0000000,
"output_cost_per_token": 0.000000,
"litellm_provider": "perplexity",
"mode": "chat"
},
"replit-code-v1.5-3b": {
"max_tokens": 4096,
"input_cost_per_token": 0.0000000,
"output_cost_per_token": 0.000000,
"litellm_provider": "perplexity",
"mode": "chat"
} }
} }