From 71e64c34cb403f1d550145f2cf370f612871ce04 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Tue, 5 Dec 2023 07:28:03 -0800 Subject: [PATCH] fix(huggingface_restapi.py): raise better exceptions for unprocessable hf responses --- litellm/llms/huggingface_restapi.py | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/litellm/llms/huggingface_restapi.py b/litellm/llms/huggingface_restapi.py index 25aa1c574..c347910f8 100644 --- a/litellm/llms/huggingface_restapi.py +++ b/litellm/llms/huggingface_restapi.py @@ -170,6 +170,11 @@ class Huggingface(BaseLLM): "content" ] = completion_response["generated_text"] # type: ignore elif task == "text-generation-inference": + if (not isinstance(completion_response, list) + or not isinstance(completion_response[0], dict) + or "generated_text" not in completion_response[0]): + raise HuggingfaceError(status_code=422, message=f"response is not in expected format - {completion_response}") + if len(completion_response[0]["generated_text"]) > 0: model_response["choices"][0]["message"][ "content"