diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 75d0ba55f3..897e9c3b26 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -936,7 +936,14 @@ "mode": "chat" }, "openrouter/mistralai/mistral-7b-instruct": { - "max_tokens": 4096, + "max_tokens": 8192, + "input_cost_per_token": 0.00000013, + "output_cost_per_token": 0.00000013, + "litellm_provider": "openrouter", + "mode": "chat" + }, + "openrouter/mistralai/mistral-7b-instruct:free": { + "max_tokens": 8192, "input_cost_per_token": 0.0, "output_cost_per_token": 0.0, "litellm_provider": "openrouter", diff --git a/litellm/router.py b/litellm/router.py index 30cb766d50..06efd4c34b 100644 --- a/litellm/router.py +++ b/litellm/router.py @@ -142,11 +142,13 @@ class Router: Router: An instance of the litellm.Router class. """ self.set_verbose = set_verbose - if self.set_verbose: + self.debug_level = debug_level + if self.set_verbose == True: if debug_level == "INFO": verbose_router_logger.setLevel(logging.INFO) elif debug_level == "DEBUG": verbose_router_logger.setLevel(logging.DEBUG) + self.deployment_names: List = ( [] ) # names of models under litellm_params. ex. azure/chatgpt-v-2 @@ -273,6 +275,16 @@ class Router: f"Intialized router with Routing strategy: {self.routing_strategy}\n" ) + def print_deployment(self, deployment: dict): + """ + returns a copy of the deployment with the api key masked + """ + _deployment_copy = copy.deepcopy(deployment) + litellm_params: dict = _deployment_copy["litellm_params"] + if "api_key" in litellm_params: + litellm_params["api_key"] = litellm_params["api_key"][:2] + "*" * 10 + return _deployment_copy + ### COMPLETION, EMBEDDING, IMG GENERATION FUNCTIONS def completion( @@ -2060,7 +2072,7 @@ class Router: verbose_router_logger.debug(f"\n selected index, {selected_index}") deployment = healthy_deployments[selected_index] verbose_router_logger.info( - f"get_available_deployment for model: {model}, Selected deployment: {deployment or deployment[0]} for model: {model}" + f"get_available_deployment for model: {model}, Selected deployment: {self.print_deployment(deployment) or deployment[0]} for model: {model}" ) return deployment or deployment[0] ############## Check if we can do a RPM/TPM based weighted pick ################# @@ -2077,7 +2089,7 @@ class Router: verbose_router_logger.debug(f"\n selected index, {selected_index}") deployment = healthy_deployments[selected_index] verbose_router_logger.info( - f"get_available_deployment for model: {model}, Selected deployment: {deployment or deployment[0]} for model: {model}" + f"get_available_deployment for model: {model}, Selected deployment: {self.print_deployment(deployment) or deployment[0]} for model: {model}" ) return deployment or deployment[0] @@ -2108,7 +2120,7 @@ class Router: ) raise ValueError("No models available.") verbose_router_logger.info( - f"get_available_deployment for model: {model}, Selected deployment: {deployment} for model: {model}" + f"get_available_deployment for model: {model}, Selected deployment: {self.print_deployment(deployment)} for model: {model}" ) return deployment