mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
(feat) redis-semantic cache
This commit is contained in:
parent
0c488a7cc7
commit
ef32a5da1b
1 changed files with 9 additions and 1 deletions
|
@ -55,7 +55,7 @@ from .integrations.litedebugger import LiteDebugger
|
||||||
from .proxy._types import KeyManagementSystem
|
from .proxy._types import KeyManagementSystem
|
||||||
from openai import OpenAIError as OriginalError
|
from openai import OpenAIError as OriginalError
|
||||||
from openai._models import BaseModel as OpenAIObject
|
from openai._models import BaseModel as OpenAIObject
|
||||||
from .caching import S3Cache
|
from .caching import S3Cache, RedisSemanticCache
|
||||||
from .exceptions import (
|
from .exceptions import (
|
||||||
AuthenticationError,
|
AuthenticationError,
|
||||||
BadRequestError,
|
BadRequestError,
|
||||||
|
@ -2535,6 +2535,14 @@ def client(original_function):
|
||||||
):
|
):
|
||||||
if len(cached_result) == 1 and cached_result[0] is None:
|
if len(cached_result) == 1 and cached_result[0] is None:
|
||||||
cached_result = None
|
cached_result = None
|
||||||
|
elif isinstance(litellm.cache.cache, RedisSemanticCache):
|
||||||
|
preset_cache_key = litellm.cache.get_cache_key(*args, **kwargs)
|
||||||
|
kwargs[
|
||||||
|
"preset_cache_key"
|
||||||
|
] = preset_cache_key # for streaming calls, we need to pass the preset_cache_key
|
||||||
|
cached_result = await litellm.cache.async_get_cache(
|
||||||
|
*args, **kwargs
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
preset_cache_key = litellm.cache.get_cache_key(*args, **kwargs)
|
preset_cache_key = litellm.cache.get_cache_key(*args, **kwargs)
|
||||||
kwargs[
|
kwargs[
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue