From 443910965886901a5c0fa4dadc2ce8f220df55bb Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Mon, 13 Nov 2023 10:29:23 -0800 Subject: [PATCH] (fix) text completion response --- litellm/utils.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/litellm/utils.py b/litellm/utils.py index c11d29341..858386f1f 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -385,6 +385,7 @@ class TextCompletionResponse(OpenAIObject): } """ def __init__(self, id=None, choices=None, created=None, model=None, usage=None, stream=False, response_ms=None, **params): + super(TextCompletionResponse, self).__init__(**params) if stream: self.object = "text_completion.chunk" self.choices = [TextChoices()] @@ -409,7 +410,7 @@ class TextCompletionResponse(OpenAIObject): else: self.usage = Usage() self._hidden_params = {} # used in case users want to access the original model response - super(TextCompletionResponse, self).__init__(**params) + def __contains__(self, key): # Define custom behavior for the 'in' operator