forked from phoenix/litellm-mirror
[Feat] Add Azure gpt-35-turbo-0301 pricing (#5790)
* add gpt-35-turbo-0301 pricing * add azure gpt-35-turbo-0613 pricing * add gpt-35-turbo-instruct-0914 pricing
This commit is contained in:
parent
d46660ea0f
commit
7e30bcc128
2 changed files with 60 additions and 0 deletions
|
@ -744,6 +744,28 @@
|
||||||
"supports_function_calling": true,
|
"supports_function_calling": true,
|
||||||
"supports_parallel_function_calling": true
|
"supports_parallel_function_calling": true
|
||||||
},
|
},
|
||||||
|
"azure/gpt-35-turbo-0613": {
|
||||||
|
"max_tokens": 4097,
|
||||||
|
"max_input_tokens": 4097,
|
||||||
|
"max_output_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000015,
|
||||||
|
"output_cost_per_token": 0.000002,
|
||||||
|
"litellm_provider": "azure",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_parallel_function_calling": true
|
||||||
|
},
|
||||||
|
"azure/gpt-35-turbo-0301": {
|
||||||
|
"max_tokens": 4097,
|
||||||
|
"max_input_tokens": 4097,
|
||||||
|
"max_output_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000002,
|
||||||
|
"output_cost_per_token": 0.000002,
|
||||||
|
"litellm_provider": "azure",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_parallel_function_calling": true
|
||||||
|
},
|
||||||
"azure/gpt-35-turbo-0125": {
|
"azure/gpt-35-turbo-0125": {
|
||||||
"max_tokens": 4096,
|
"max_tokens": 4096,
|
||||||
"max_input_tokens": 16384,
|
"max_input_tokens": 16384,
|
||||||
|
@ -790,6 +812,14 @@
|
||||||
"litellm_provider": "text-completion-openai",
|
"litellm_provider": "text-completion-openai",
|
||||||
"mode": "completion"
|
"mode": "completion"
|
||||||
},
|
},
|
||||||
|
"azure/gpt-35-turbo-instruct-0914": {
|
||||||
|
"max_tokens": 4097,
|
||||||
|
"max_input_tokens": 4097,
|
||||||
|
"input_cost_per_token": 0.0000015,
|
||||||
|
"output_cost_per_token": 0.000002,
|
||||||
|
"litellm_provider": "text-completion-openai",
|
||||||
|
"mode": "completion"
|
||||||
|
},
|
||||||
"azure/mistral-large-latest": {
|
"azure/mistral-large-latest": {
|
||||||
"max_tokens": 32000,
|
"max_tokens": 32000,
|
||||||
"max_input_tokens": 32000,
|
"max_input_tokens": 32000,
|
||||||
|
|
|
@ -744,6 +744,28 @@
|
||||||
"supports_function_calling": true,
|
"supports_function_calling": true,
|
||||||
"supports_parallel_function_calling": true
|
"supports_parallel_function_calling": true
|
||||||
},
|
},
|
||||||
|
"azure/gpt-35-turbo-0613": {
|
||||||
|
"max_tokens": 4097,
|
||||||
|
"max_input_tokens": 4097,
|
||||||
|
"max_output_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000015,
|
||||||
|
"output_cost_per_token": 0.000002,
|
||||||
|
"litellm_provider": "azure",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_parallel_function_calling": true
|
||||||
|
},
|
||||||
|
"azure/gpt-35-turbo-0301": {
|
||||||
|
"max_tokens": 4097,
|
||||||
|
"max_input_tokens": 4097,
|
||||||
|
"max_output_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000002,
|
||||||
|
"output_cost_per_token": 0.000002,
|
||||||
|
"litellm_provider": "azure",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_parallel_function_calling": true
|
||||||
|
},
|
||||||
"azure/gpt-35-turbo-0125": {
|
"azure/gpt-35-turbo-0125": {
|
||||||
"max_tokens": 4096,
|
"max_tokens": 4096,
|
||||||
"max_input_tokens": 16384,
|
"max_input_tokens": 16384,
|
||||||
|
@ -790,6 +812,14 @@
|
||||||
"litellm_provider": "text-completion-openai",
|
"litellm_provider": "text-completion-openai",
|
||||||
"mode": "completion"
|
"mode": "completion"
|
||||||
},
|
},
|
||||||
|
"azure/gpt-35-turbo-instruct-0914": {
|
||||||
|
"max_tokens": 4097,
|
||||||
|
"max_input_tokens": 4097,
|
||||||
|
"input_cost_per_token": 0.0000015,
|
||||||
|
"output_cost_per_token": 0.000002,
|
||||||
|
"litellm_provider": "text-completion-openai",
|
||||||
|
"mode": "completion"
|
||||||
|
},
|
||||||
"azure/mistral-large-latest": {
|
"azure/mistral-large-latest": {
|
||||||
"max_tokens": 32000,
|
"max_tokens": 32000,
|
||||||
"max_input_tokens": 32000,
|
"max_input_tokens": 32000,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue