mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
fix(init.py): rename feature_flag
This commit is contained in:
parent
3c4c78a71f
commit
a9fdfb5a99
3 changed files with 4 additions and 4 deletions
|
@ -146,7 +146,7 @@ return_response_headers: bool = (
|
||||||
)
|
)
|
||||||
##################
|
##################
|
||||||
logging: bool = True
|
logging: bool = True
|
||||||
enable_caching_on_optional_params: bool = (
|
enable_caching_on_provider_specific_optional_params: bool = (
|
||||||
False # feature-flag for caching on optional params - e.g. 'top_k'
|
False # feature-flag for caching on optional params - e.g. 'top_k'
|
||||||
)
|
)
|
||||||
caching: bool = (
|
caching: bool = (
|
||||||
|
|
|
@ -1903,7 +1903,7 @@ class Cache:
|
||||||
param not in litellm_param_kwargs
|
param not in litellm_param_kwargs
|
||||||
): # check if user passed in optional param - e.g. top_k
|
): # check if user passed in optional param - e.g. top_k
|
||||||
if (
|
if (
|
||||||
litellm.enable_caching_on_optional_params is True
|
litellm.enable_caching_on_provider_specific_optional_params is True
|
||||||
): # feature flagged for now
|
): # feature flagged for now
|
||||||
if kwargs[param] is None:
|
if kwargs[param] is None:
|
||||||
continue # ignore None params
|
continue # ignore None params
|
||||||
|
|
|
@ -302,7 +302,7 @@ def test_caching_with_models_v2():
|
||||||
|
|
||||||
|
|
||||||
def test_caching_with_optional_params():
|
def test_caching_with_optional_params():
|
||||||
litellm.enable_caching_on_optional_params = True
|
litellm.enable_caching_on_provider_specific_optional_params = True
|
||||||
messages = [
|
messages = [
|
||||||
{"role": "user", "content": "who is ishaan CTO of litellm from litellm 2023"}
|
{"role": "user", "content": "who is ishaan CTO of litellm from litellm 2023"}
|
||||||
]
|
]
|
||||||
|
@ -352,7 +352,7 @@ def test_caching_with_optional_params():
|
||||||
print(f"response1: {response1}")
|
print(f"response1: {response1}")
|
||||||
print(f"response2: {response2}")
|
print(f"response2: {response2}")
|
||||||
pytest.fail(f"Error occurred:")
|
pytest.fail(f"Error occurred:")
|
||||||
litellm.enable_caching_on_optional_params = False
|
litellm.enable_caching_on_provider_specific_optional_params = False
|
||||||
|
|
||||||
|
|
||||||
embedding_large_text = (
|
embedding_large_text = (
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue