fix route_llm_request

This commit is contained in:
Ishaan Jaff 2024-11-27 15:35:19 -08:00
parent d717fd8fcb
commit 2c84b19550

View file

@ -86,7 +86,6 @@ async def route_request(
else:
models = [model.strip() for model in data.pop("model").split(",")]
return llm_router.abatch_completion(models=models, **data)
elif llm_router is not None:
if (
data["model"] in router_model_names
@ -113,6 +112,9 @@ async def route_request(
or len(llm_router.pattern_router.patterns) > 0
):
return getattr(llm_router, f"{route_type}")(**data)
elif route_type == "amoderation":
# moderation endpoint does not require `model` parameter
return getattr(llm_router, f"{route_type}")(**data)
elif user_model is not None:
return getattr(litellm, f"{route_type}")(**data)