mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix(utils.py): handle scenario where model="azure/*" and custom_llm_provider="azure"
Fixes https://github.com/BerriAI/litellm/issues/4912
This commit is contained in:
parent
3ee8ae231c
commit
5d96ff6694
5 changed files with 23 additions and 21 deletions
|
@ -472,11 +472,10 @@ class _PROXY_MaxParallelRequestsHandler(CustomLogger):
|
|||
|
||||
async def async_log_failure_event(self, kwargs, response_obj, start_time, end_time):
|
||||
try:
|
||||
self.print_verbose(f"Inside Max Parallel Request Failure Hook")
|
||||
global_max_parallel_requests = (
|
||||
kwargs["litellm_params"]
|
||||
.get("metadata", {})
|
||||
.get("global_max_parallel_requests", None)
|
||||
self.print_verbose("Inside Max Parallel Request Failure Hook")
|
||||
_metadata = kwargs["litellm_params"].get("metadata", {}) or {}
|
||||
global_max_parallel_requests = _metadata.get(
|
||||
"global_max_parallel_requests", None
|
||||
)
|
||||
user_api_key = (
|
||||
kwargs["litellm_params"].get("metadata", {}).get("user_api_key", None)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue