From 2c84b19550b84e4e62bacc362b737223807f86d8 Mon Sep 17 00:00:00 2001 From: Ishaan Jaff Date: Wed, 27 Nov 2024 15:35:19 -0800 Subject: [PATCH] fix route_llm_request --- litellm/proxy/route_llm_request.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/litellm/proxy/route_llm_request.py b/litellm/proxy/route_llm_request.py index 3c5c8b3b4..ec9850eeb 100644 --- a/litellm/proxy/route_llm_request.py +++ b/litellm/proxy/route_llm_request.py @@ -86,7 +86,6 @@ async def route_request( else: models = [model.strip() for model in data.pop("model").split(",")] return llm_router.abatch_completion(models=models, **data) - elif llm_router is not None: if ( data["model"] in router_model_names @@ -113,6 +112,9 @@ async def route_request( or len(llm_router.pattern_router.patterns) > 0 ): return getattr(llm_router, f"{route_type}")(**data) + elif route_type == "amoderation": + # moderation endpoint does not require `model` parameter + return getattr(llm_router, f"{route_type}")(**data) elif user_model is not None: return getattr(litellm, f"{route_type}")(**data)