forked from phoenix/litellm-mirror
fix linting errors
This commit is contained in:
parent
1e1bce4594
commit
66211b42db
2 changed files with 5 additions and 3 deletions
|
@ -2609,7 +2609,7 @@ class OpenAIBatchesAPI(BaseLLM):
|
|||
limit: Optional[int] = None,
|
||||
):
|
||||
verbose_logger.debug("listing batches, after= %s, limit= %s", after, limit)
|
||||
response = await openai_client.batches.list(after=after, limit=limit)
|
||||
response = await openai_client.batches.list(after=after, limit=limit) # type: ignore
|
||||
return response
|
||||
|
||||
def list_batches(
|
||||
|
@ -2646,7 +2646,7 @@ class OpenAIBatchesAPI(BaseLLM):
|
|||
return self.alist_batches( # type: ignore
|
||||
openai_client=openai_client, after=after, limit=limit
|
||||
)
|
||||
response = openai_client.batches.list(after=after, limit=limit)
|
||||
response = openai_client.batches.list(after=after, limit=limit) # type: ignore
|
||||
return response
|
||||
|
||||
|
||||
|
|
|
@ -5052,7 +5052,9 @@ async def list_batches(
|
|||
return response
|
||||
except Exception as e:
|
||||
await proxy_logging_obj.post_call_failure_hook(
|
||||
user_api_key_dict=user_api_key_dict, original_exception=e, request_data=data
|
||||
user_api_key_dict=user_api_key_dict,
|
||||
original_exception=e,
|
||||
request_data={"after": after, "limit": limit},
|
||||
)
|
||||
verbose_proxy_logger.error(
|
||||
"litellm.proxy.proxy_server.retrieve_batch(): Exception occured - {}".format(
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue