mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 19:24:27 +00:00
fix(proxy_server.py): ignore cache if value is false
This commit is contained in:
parent
6d7739aa6d
commit
8619499853
1 changed files with 2 additions and 0 deletions
|
@ -1727,6 +1727,8 @@ class ProxyConfig:
|
||||||
print( # noqa
|
print( # noqa
|
||||||
f"{blue_color_code}Set Cache on LiteLLM Proxy: {vars(litellm.cache.cache)}{reset_color_code}"
|
f"{blue_color_code}Set Cache on LiteLLM Proxy: {vars(litellm.cache.cache)}{reset_color_code}"
|
||||||
)
|
)
|
||||||
|
elif key == "cache" and value == False:
|
||||||
|
pass
|
||||||
elif key == "callbacks":
|
elif key == "callbacks":
|
||||||
if isinstance(value, list):
|
if isinstance(value, list):
|
||||||
imported_list: List[Any] = []
|
imported_list: List[Any] = []
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue