From da4441b1fc7eb026a3c9f16089a241406cc68b51 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Fri, 9 Feb 2024 16:47:01 -0800 Subject: [PATCH] (feat) support azure deployments for embeddings --- litellm/proxy/proxy_server.py | 8 ++++++++ 1 file changed, 8 insertions(+) diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index a2d03486da..693afbe1c0 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -2433,8 +2433,15 @@ async def chat_completion( response_class=ORJSONResponse, tags=["embeddings"], ) +@router.post( + "/openai/deployments/{model:path}/embeddings", + dependencies=[Depends(user_api_key_auth)], + response_class=ORJSONResponse, + tags=["embeddings"], +) # azure compatible endpoint async def embeddings( request: Request, + model: Optional[str] = None, user_api_key_dict: UserAPIKeyAuth = Depends(user_api_key_auth), background_tasks: BackgroundTasks = BackgroundTasks(), ): @@ -2458,6 +2465,7 @@ async def embeddings( data["model"] = ( general_settings.get("embedding_model", None) # server default or user_model # model name passed via cli args + or model # for azure deployments or data["model"] # default passed in http request ) if user_model: