From 363d74cf7d585ecf5567518c66515ed276e3b0ef Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Tue, 21 Nov 2023 19:31:50 -0800 Subject: [PATCH] (fix) proxy server set model list through headers --- litellm/proxy/proxy_server.py | 11 +++-------- 1 file changed, 3 insertions(+), 8 deletions(-) diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index e93527eae..8cf1dd5e5 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -176,15 +176,10 @@ async def user_api_key_auth(request: Request): litellm.model_alias_map = valid_token.aliases config = valid_token.config if config != {}: - global llm_router + global llm_model_list model_list = config.get("model_list", []) - if llm_router == None: - llm_router = litellm.Router( - model_list=model_list - ) - else: - llm_router.model_list = model_list - print("\n new llm router model list", llm_router.model_list) + llm_model_list = model_list + print("\n new llm router model list", llm_model_list) if len(valid_token.models) == 0: # assume an empty model list means all models are allowed to be called return else: