From 66211b42db0854d0777bafbe91efa20e11e2a9de Mon Sep 17 00:00:00 2001 From: Ishaan Jaff Date: Tue, 30 Jul 2024 12:51:39 -0700 Subject: [PATCH] fix linting errors --- litellm/llms/openai.py | 4 ++-- litellm/proxy/proxy_server.py | 4 +++- 2 files changed, 5 insertions(+), 3 deletions(-) diff --git a/litellm/llms/openai.py b/litellm/llms/openai.py index fab3ff26d..d2ba7ac13 100644 --- a/litellm/llms/openai.py +++ b/litellm/llms/openai.py @@ -2609,7 +2609,7 @@ class OpenAIBatchesAPI(BaseLLM): limit: Optional[int] = None, ): verbose_logger.debug("listing batches, after= %s, limit= %s", after, limit) - response = await openai_client.batches.list(after=after, limit=limit) + response = await openai_client.batches.list(after=after, limit=limit) # type: ignore return response def list_batches( @@ -2646,7 +2646,7 @@ class OpenAIBatchesAPI(BaseLLM): return self.alist_batches( # type: ignore openai_client=openai_client, after=after, limit=limit ) - response = openai_client.batches.list(after=after, limit=limit) + response = openai_client.batches.list(after=after, limit=limit) # type: ignore return response diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index cbb62289b..5a2970df5 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -5052,7 +5052,9 @@ async def list_batches( return response except Exception as e: await proxy_logging_obj.post_call_failure_hook( - user_api_key_dict=user_api_key_dict, original_exception=e, request_data=data + user_api_key_dict=user_api_key_dict, + original_exception=e, + request_data={"after": after, "limit": limit}, ) verbose_proxy_logger.error( "litellm.proxy.proxy_server.retrieve_batch(): Exception occured - {}".format(