From aa6f5e13f3439c555cf8be0f95dd6191a8824628 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Thu, 21 Sep 2023 15:44:13 -0700 Subject: [PATCH] response_ms fix --- litellm/utils.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/litellm/utils.py b/litellm/utils.py index 0bc3f6a427..2590df7936 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -142,9 +142,9 @@ class ModelResponse(OpenAIObject): else: self.created = created if response_ms: - self.response_ms = response_ms + self._response_ms = response_ms else: - self.response_ms = None + self._response_ms = None self.model = model self.usage = ( usage @@ -639,7 +639,7 @@ def client(original_function): ) # don't interrupt execution of main thread my_thread.start() # RETURN RESULT - result.response_ms = (end_time - start_time).total_seconds() * 1000 # return response latency in ms like openai + result._response_ms = (end_time - start_time).total_seconds() * 1000 # return response latency in ms like openai return result except Exception as e: traceback_exception = traceback.format_exc()