mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
fix feature flag logic
This commit is contained in:
parent
db22da0dbc
commit
99fd2228ba
1 changed files with 2 additions and 3 deletions
|
@ -2558,9 +2558,8 @@ class Router:
|
|||
self.set_client(model=deployment.to_json(exclude_none=True))
|
||||
|
||||
# set region (if azure model)
|
||||
_auto_infer_region = os.environ.get("AUTO_INFER_REGION", "true")
|
||||
_auto_infer_region_value = bool(_auto_infer_region)
|
||||
if _auto_infer_region_value == True:
|
||||
_auto_infer_region = os.environ.get("AUTO_INFER_REGION", False)
|
||||
if _auto_infer_region == True:
|
||||
"""
|
||||
Hiding behind a feature flag
|
||||
When there is a large amount of LLM deployments this makes startup times blow up
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue