mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
(feat) add perplexity llms
This commit is contained in:
parent
0b94e968e1
commit
1bd41ab636
2 changed files with 40 additions and 0 deletions
|
@ -100,6 +100,7 @@ nlp_cloud_models: List = []
|
||||||
aleph_alpha_models: List = []
|
aleph_alpha_models: List = []
|
||||||
bedrock_models: List = []
|
bedrock_models: List = []
|
||||||
deepinfra_models: List = []
|
deepinfra_models: List = []
|
||||||
|
perplexity_models: List = []
|
||||||
for key, value in model_cost.items():
|
for key, value in model_cost.items():
|
||||||
if value.get('litellm_provider') == 'openai':
|
if value.get('litellm_provider') == 'openai':
|
||||||
open_ai_chat_completion_models.append(key)
|
open_ai_chat_completion_models.append(key)
|
||||||
|
@ -130,6 +131,8 @@ for key, value in model_cost.items():
|
||||||
bedrock_models.append(key)
|
bedrock_models.append(key)
|
||||||
elif value.get('litellm_provider') == 'deepinfra':
|
elif value.get('litellm_provider') == 'deepinfra':
|
||||||
deepinfra_models.append(key)
|
deepinfra_models.append(key)
|
||||||
|
elif value.get('litellm_provider') == 'perplexity':
|
||||||
|
perplexity_models.append(key)
|
||||||
|
|
||||||
# known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary
|
# known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary
|
||||||
openai_compatible_endpoints: List = [
|
openai_compatible_endpoints: List = [
|
||||||
|
@ -234,6 +237,7 @@ model_list = (
|
||||||
+ ollama_models
|
+ ollama_models
|
||||||
+ bedrock_models
|
+ bedrock_models
|
||||||
+ deepinfra_models
|
+ deepinfra_models
|
||||||
|
+ perplexity_models
|
||||||
)
|
)
|
||||||
|
|
||||||
provider_list: List = [
|
provider_list: List = [
|
||||||
|
@ -277,6 +281,7 @@ models_by_provider: dict = {
|
||||||
"petals": petals_models,
|
"petals": petals_models,
|
||||||
"ollama": ollama_models,
|
"ollama": ollama_models,
|
||||||
"deepinfra": deepinfra_models,
|
"deepinfra": deepinfra_models,
|
||||||
|
"perplexity": perplexity_models,
|
||||||
}
|
}
|
||||||
|
|
||||||
# mapping for those models which have larger equivalents
|
# mapping for those models which have larger equivalents
|
||||||
|
|
|
@ -612,5 +612,40 @@
|
||||||
"output_cost_per_token": 0.00000095,
|
"output_cost_per_token": 0.00000095,
|
||||||
"litellm_provider": "deepinfra",
|
"litellm_provider": "deepinfra",
|
||||||
"mode": "chat"
|
"mode": "chat"
|
||||||
|
},
|
||||||
|
"codellama-34b-instruct": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000000,
|
||||||
|
"output_cost_per_token": 0.000000,
|
||||||
|
"litellm_provider": "perplexity",
|
||||||
|
"mode": "chat"
|
||||||
|
},
|
||||||
|
"llama-2-13b-chat": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000000,
|
||||||
|
"output_cost_per_token": 0.000000,
|
||||||
|
"litellm_provider": "perplexity",
|
||||||
|
"mode": "chat"
|
||||||
|
},
|
||||||
|
"llama-2-70b-chat": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000000,
|
||||||
|
"output_cost_per_token": 0.000000,
|
||||||
|
"litellm_provider": "perplexity",
|
||||||
|
"mode": "chat"
|
||||||
|
},
|
||||||
|
"mistral-7b-instruct": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000000,
|
||||||
|
"output_cost_per_token": 0.000000,
|
||||||
|
"litellm_provider": "perplexity",
|
||||||
|
"mode": "chat"
|
||||||
|
},
|
||||||
|
"replit-code-v1.5-3b": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000000,
|
||||||
|
"output_cost_per_token": 0.000000,
|
||||||
|
"litellm_provider": "perplexity",
|
||||||
|
"mode": "chat"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue