fix(proxy_server.py): support model info augmenting for azure models

This commit is contained in:
Krrish Dholakia 2023-12-02 21:33:47 -08:00
parent ecddb852a2
commit add4dfc528
2 changed files with 11 additions and 1 deletions

View file

@ -1093,8 +1093,11 @@ async def model_info(request: Request):
data = copy.deepcopy(model["litellm_params"])
data["messages"] = [{"role": "user", "content": "Hey, how's it going?"}]
data["max_tokens"] = 10
print(f"data going to litellm acompletion: {data}")
response = await litellm.acompletion(**data)
litellm_model_info = litellm.model_cost.get(response["model"], {})
response_model = response["model"]
print(f"response model: {response_model}; response - {response}")
litellm_model_info = litellm.get_model_info(response_model)
model_info = model.get("model_info", {})
for k, v in litellm_model_info.items():
if k not in model_info: