fix(parallel_request_limiter.py): decrement count for failed llm calls

https://github.com/BerriAI/litellm/issues/1477
This commit is contained in:
Krrish Dholakia 2024-01-18 12:42:14 -08:00
parent 37e6c6a59f
commit 1ea3833ef7
3 changed files with 350 additions and 27 deletions

View file

@ -1102,7 +1102,7 @@ async def generate_key_helper_fn(
}
if prisma_client is not None:
## CREATE USER (If necessary)
verbose_proxy_logger.debug(f"CustomDBClient: Creating User={user_data}")
verbose_proxy_logger.debug(f"prisma_client: Creating User={user_data}")
user_row = await prisma_client.insert_data(
data=user_data, table_name="user"
)
@ -1111,7 +1111,7 @@ async def generate_key_helper_fn(
if len(user_row.models) > 0 and len(key_data["models"]) == 0: # type: ignore
key_data["models"] = user_row.models
## CREATE KEY
verbose_proxy_logger.debug(f"CustomDBClient: Creating Key={key_data}")
verbose_proxy_logger.debug(f"prisma_client: Creating Key={key_data}")
await prisma_client.insert_data(data=key_data, table_name="key")
elif custom_db_client is not None:
## CREATE USER (If necessary)