mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
Update model_prices_and_context_window.json
1. Remove duplicate key-value pairs; 2. Add ollama model context "ollama/qwen-2.5-coder"
This commit is contained in:
parent
d4caaae1be
commit
ddb38ae42a
1 changed files with 20 additions and 18 deletions
|
@ -8634,6 +8634,26 @@
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"supports_tool_choice": true
|
"supports_tool_choice": true
|
||||||
},
|
},
|
||||||
|
"ollama/qwen-coder-14b": {
|
||||||
|
"max_tokens": 8192,
|
||||||
|
"max_input_tokens": 8192,
|
||||||
|
"max_output_tokens": 8192,
|
||||||
|
"input_cost_per_token": 0.0,
|
||||||
|
"output_cost_per_token": 0.0,
|
||||||
|
"litellm_provider": "ollama",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true
|
||||||
|
},
|
||||||
|
"ollama/qwen-coder-32b": {
|
||||||
|
"max_tokens": 32768,
|
||||||
|
"max_input_tokens": 32768,
|
||||||
|
"max_output_tokens": 8192,
|
||||||
|
"input_cost_per_token": 0.0,
|
||||||
|
"output_cost_per_token": 0.0,
|
||||||
|
"litellm_provider": "ollama",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true
|
||||||
|
},
|
||||||
"ollama/codegemma": {
|
"ollama/codegemma": {
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"max_input_tokens": 8192,
|
"max_input_tokens": 8192,
|
||||||
|
@ -9179,24 +9199,6 @@
|
||||||
"litellm_provider": "perplexity",
|
"litellm_provider": "perplexity",
|
||||||
"mode": "chat"
|
"mode": "chat"
|
||||||
},
|
},
|
||||||
"perplexity/sonar": {
|
|
||||||
"max_tokens": 127072,
|
|
||||||
"max_input_tokens": 127072,
|
|
||||||
"max_output_tokens": 127072,
|
|
||||||
"input_cost_per_token": 0.000001,
|
|
||||||
"output_cost_per_token": 0.000001,
|
|
||||||
"litellm_provider": "perplexity",
|
|
||||||
"mode": "chat"
|
|
||||||
},
|
|
||||||
"perplexity/sonar-pro": {
|
|
||||||
"max_tokens": 200000,
|
|
||||||
"max_input_tokens": 200000,
|
|
||||||
"max_output_tokens": 8096,
|
|
||||||
"input_cost_per_token": 0.000003,
|
|
||||||
"output_cost_per_token": 0.000015,
|
|
||||||
"litellm_provider": "perplexity",
|
|
||||||
"mode": "chat"
|
|
||||||
},
|
|
||||||
"perplexity/pplx-7b-chat": {
|
"perplexity/pplx-7b-chat": {
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"max_input_tokens": 8192,
|
"max_input_tokens": 8192,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue