mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
feat(proxy_server.py): move credential list to being a top-level param
This commit is contained in:
parent
1e89fdec99
commit
236e68910c
4 changed files with 14 additions and 2 deletions
|
@ -287,7 +287,7 @@ from litellm.types.llms.openai import HttpxBinaryResponseContent
|
|||
from litellm.types.router import DeploymentTypedDict
|
||||
from litellm.types.router import ModelInfo as RouterModelInfo
|
||||
from litellm.types.router import RouterGeneralSettings, updateDeployment
|
||||
from litellm.types.utils import CustomHuggingfaceTokenizer
|
||||
from litellm.types.utils import CredentialItem, CustomHuggingfaceTokenizer
|
||||
from litellm.types.utils import ModelInfo as ModelMapInfo
|
||||
from litellm.types.utils import RawRequestTypedDict, StandardLoggingPayload
|
||||
from litellm.utils import _add_custom_logger_callback_to_specific_event
|
||||
|
@ -2184,6 +2184,9 @@ class ProxyConfig:
|
|||
init_guardrails_v2(
|
||||
all_guardrails=guardrails_v2, config_file_path=config_file_path
|
||||
)
|
||||
|
||||
## CREDENTIALS
|
||||
litellm.credential_list = config.get("credential_list")
|
||||
return router, router.get_model_list(), general_settings
|
||||
|
||||
def _load_alerting_settings(self, general_settings: dict):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue