forked from phoenix/litellm-mirror
(feat) add perplexity llms
This commit is contained in:
parent
0b94e968e1
commit
1bd41ab636
2 changed files with 40 additions and 0 deletions
|
@ -100,6 +100,7 @@ nlp_cloud_models: List = []
|
|||
aleph_alpha_models: List = []
|
||||
bedrock_models: List = []
|
||||
deepinfra_models: List = []
|
||||
perplexity_models: List = []
|
||||
for key, value in model_cost.items():
|
||||
if value.get('litellm_provider') == 'openai':
|
||||
open_ai_chat_completion_models.append(key)
|
||||
|
@ -130,6 +131,8 @@ for key, value in model_cost.items():
|
|||
bedrock_models.append(key)
|
||||
elif value.get('litellm_provider') == 'deepinfra':
|
||||
deepinfra_models.append(key)
|
||||
elif value.get('litellm_provider') == 'perplexity':
|
||||
perplexity_models.append(key)
|
||||
|
||||
# known openai compatible endpoints - we'll eventually move this list to the model_prices_and_context_window.json dictionary
|
||||
openai_compatible_endpoints: List = [
|
||||
|
@ -234,6 +237,7 @@ model_list = (
|
|||
+ ollama_models
|
||||
+ bedrock_models
|
||||
+ deepinfra_models
|
||||
+ perplexity_models
|
||||
)
|
||||
|
||||
provider_list: List = [
|
||||
|
@ -277,6 +281,7 @@ models_by_provider: dict = {
|
|||
"petals": petals_models,
|
||||
"ollama": ollama_models,
|
||||
"deepinfra": deepinfra_models,
|
||||
"perplexity": perplexity_models,
|
||||
}
|
||||
|
||||
# mapping for those models which have larger equivalents
|
||||
|
|
|
@ -612,5 +612,40 @@
|
|||
"output_cost_per_token": 0.00000095,
|
||||
"litellm_provider": "deepinfra",
|
||||
"mode": "chat"
|
||||
},
|
||||
"codellama-34b-instruct": {
|
||||
"max_tokens": 4096,
|
||||
"input_cost_per_token": 0.0000000,
|
||||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
},
|
||||
"llama-2-13b-chat": {
|
||||
"max_tokens": 4096,
|
||||
"input_cost_per_token": 0.0000000,
|
||||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
},
|
||||
"llama-2-70b-chat": {
|
||||
"max_tokens": 4096,
|
||||
"input_cost_per_token": 0.0000000,
|
||||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
},
|
||||
"mistral-7b-instruct": {
|
||||
"max_tokens": 4096,
|
||||
"input_cost_per_token": 0.0000000,
|
||||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
},
|
||||
"replit-code-v1.5-3b": {
|
||||
"max_tokens": 4096,
|
||||
"input_cost_per_token": 0.0000000,
|
||||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "perplexity",
|
||||
"mode": "chat"
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue