forked from phoenix/litellm-mirror
docs fix
This commit is contained in:
parent
d533acd24a
commit
6e6d38841f
2 changed files with 5 additions and 5 deletions
|
@ -660,7 +660,7 @@ general_settings:
|
|||
| enable_jwt_auth | boolean | allow proxy admin to auth in via jwt tokens with 'litellm_proxy_admin' in claims. [Doc on JWT Tokens](token_auth) |
|
||||
| enforce_user_param | boolean | If true, requires all OpenAI endpoint requests to have a 'user' param. [Doc on call hooks](call_hooks)|
|
||||
| allowed_routes | array of strings | List of allowed proxy API routes a user can access [Doc on controlling allowed routes](enterprise#control-available-public-private-routes)|
|
||||
| key_management_system | string | Specifies the key management system. [Doc Secret Managers](secret) |
|
||||
| key_management_system | string | Specifies the key management system. [Doc Secret Managers](../secret) |
|
||||
| master_key | string | The master key for the proxy [Set up Virtual Keys](virtual_keys) |
|
||||
| database_url | string | The URL for the database connection [Set up Virtual Keys](virtual_keys) |
|
||||
| database_connection_pool_limit | integer | The limit for database connection pool [Setting DB Connection Pool limit](#configure-db-pool-limits--connection-timeouts) |
|
||||
|
|
|
@ -3604,7 +3604,7 @@
|
|||
"max_output_tokens": 65536,
|
||||
"input_cost_per_token": 0.000003,
|
||||
"output_cost_per_token": 0.000012,
|
||||
"litellm_provider": "openai",
|
||||
"litellm_provider": "openrouter",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true,
|
||||
|
@ -3616,7 +3616,7 @@
|
|||
"max_output_tokens": 65536,
|
||||
"input_cost_per_token": 0.000003,
|
||||
"output_cost_per_token": 0.000012,
|
||||
"litellm_provider": "openai",
|
||||
"litellm_provider": "openrouter",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true,
|
||||
|
@ -3628,7 +3628,7 @@
|
|||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000015,
|
||||
"output_cost_per_token": 0.000060,
|
||||
"litellm_provider": "openai",
|
||||
"litellm_provider": "openrouter",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true,
|
||||
|
@ -3640,7 +3640,7 @@
|
|||
"max_output_tokens": 32768,
|
||||
"input_cost_per_token": 0.000015,
|
||||
"output_cost_per_token": 0.000060,
|
||||
"litellm_provider": "openai",
|
||||
"litellm_provider": "openrouter",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_parallel_function_calling": true,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue