mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
cache improvements
This commit is contained in:
parent
43fea52d7c
commit
1799f1bfe9
4 changed files with 15 additions and 15 deletions
|
@ -30,9 +30,9 @@ togetherai_api_key: Optional[str] = None
|
|||
baseten_key: Optional[str] = None
|
||||
use_client = False
|
||||
logging = True
|
||||
caching = False
|
||||
cache: Optional[Cache] = None # set to litellm.caching Cache() object
|
||||
caching_with_models = False # if you want the caching key to be model + prompt
|
||||
caching = False # deprecated son
|
||||
caching_with_models = False # if you want the caching key to be model + prompt # deprecated soon
|
||||
cache: Optional[Cache] = None # cache object
|
||||
model_alias_map: Dict[str, str] = {}
|
||||
model_cost = {
|
||||
"babbage-002": {
|
||||
|
|
|
@ -75,7 +75,6 @@ class Cache():
|
|||
if cache_key is not None:
|
||||
self.cache.set_cache(cache_key, result)
|
||||
except:
|
||||
|
||||
pass
|
||||
|
||||
|
||||
|
@ -83,6 +82,3 @@ class Cache():
|
|||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
|
|
@ -29,13 +29,11 @@ def test_caching():
|
|||
if response2 != response1:
|
||||
print(f"response1: {response1}")
|
||||
print(f"response2: {response2}")
|
||||
pytest.fail(f"Error occurred: {e}")
|
||||
pytest.fail(f"Error occurred: responses are not equal")
|
||||
except Exception as e:
|
||||
litellm.caching = False
|
||||
print(f"error occurred: {traceback.format_exc()}")
|
||||
pytest.fail(f"Error occurred: {e}")
|
||||
|
||||
|
||||
def test_caching_with_models():
|
||||
litellm.caching_with_models = True
|
||||
response1 = completion(model="gpt-3.5-turbo", messages=messages)
|
||||
|
|
|
@ -26,6 +26,7 @@ from .exceptions import (
|
|||
OpenAIError,
|
||||
)
|
||||
from typing import List, Dict, Union, Optional
|
||||
from .caching import Cache
|
||||
|
||||
|
||||
####### ENVIRONMENT VARIABLES ####################
|
||||
|
@ -402,11 +403,16 @@ def client(original_function):
|
|||
kwargs["litellm_call_id"] = litellm_call_id
|
||||
start_time = datetime.datetime.now()
|
||||
# [OPTIONAL] CHECK CACHE
|
||||
if (litellm.caching or litellm.caching_with_models or litellm.cache != None) and (
|
||||
cached_result := litellm.cache.get_cache(*args, **kwargs)
|
||||
) is not None:
|
||||
result = cached_result
|
||||
return result
|
||||
# remove this after deprecating litellm.caching
|
||||
if (litellm.caching or litellm.caching_with_models) and litellm.cache is None:
|
||||
litellm.cache = Cache()
|
||||
|
||||
# checking cache
|
||||
if (litellm.cache != None or litellm.caching or litellm.caching_with_models):
|
||||
cached_result = litellm.cache.get_cache(*args, **kwargs)
|
||||
if cached_result != None:
|
||||
return cached_result
|
||||
|
||||
# MODEL CALL
|
||||
result = original_function(*args, **kwargs)
|
||||
if "stream" in kwargs and kwargs["stream"] == True:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue