From b51371952b4ea91df3334f1233122527dd66e7d1 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Wed, 3 Jan 2024 15:33:18 +0530 Subject: [PATCH] fix(proxy_server.py): handle base case for /model/info --- litellm/proxy/proxy_server.py | 10 +++++++--- 1 file changed, 7 insertions(+), 3 deletions(-) diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index c5249943d..59b8479d1 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -128,7 +128,7 @@ user_temperature = None user_telemetry = True user_config = None user_headers = None -user_config_file_path = f"config_{time.time()}.yaml" +user_config_file_path = f"config_{int(time.time())}.yaml" local_logging = True # writes logs to a local api_log.json file for debugging experimental = False #### GLOBAL VARIABLES #### @@ -1936,8 +1936,12 @@ async def add_new_model(model_params: ModelParams): async def model_info_v1(request: Request): global llm_model_list, general_settings, user_config_file_path # Load existing config - with open(f"{user_config_file_path}", "r") as config_file: - config = yaml.safe_load(config_file) + if os.path.exists(f"{user_config_file_path}"): + with open(f"{user_config_file_path}", "r") as config_file: + config = yaml.safe_load(config_file) + else: + config = {"model_list": []} # handle base case + all_models = config["model_list"] for model in all_models: # provided model_info in config.yaml