diff --git a/litellm/proxy/proxy_config.yaml b/litellm/proxy/proxy_config.yaml index 4ebc3060eb..1c096aac71 100644 --- a/litellm/proxy/proxy_config.yaml +++ b/litellm/proxy/proxy_config.yaml @@ -40,14 +40,11 @@ model_list: model_info: mode: completion - litellm_settings: # setting callback class # callbacks: custom_callbacks.proxy_handler_instance # sets litellm.callbacks = [proxy_handler_instance] + model_group_alias_map: {"gpt-4": "openai-gpt-3.5"} # all requests with gpt-4 model_name, get sent to openai-gpt-3.5 - # setting a callback function for success and failure - success_callback: [custom_callbacks.async_on_succes_logger] - failure_callback: [custom_callbacks.async_on_fail_logger] general_settings: # otel: True # OpenTelemetry Logger diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index 08918925b0..256e77444a 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -944,6 +944,8 @@ async def chat_completion(request: Request, model: Optional[str] = None, user_ap response = await llm_router.acompletion(**data) elif llm_router is not None and data["model"] in llm_router.deployment_names: # model in router deployments, calling a specific deployment on the router response = await llm_router.acompletion(**data, specific_deployment = True) + elif llm_router is not None and data["model"] in litellm.model_group_alias_map: # model set in model_group_alias_map + response = await llm_router.acompletion(**data) else: # router is not set response = await litellm.acompletion(**data) if 'stream' in data and data['stream'] == True: # use generate_responses to stream responses