(feat) working semantic cache on proxy

This commit is contained in:
ishaan-jaff 2024-02-06 08:55:25 -08:00
parent a3b1e3bc84
commit 70a895329e

View file

@ -73,6 +73,11 @@ litellm_settings:
max_budget: 1.5000
models: ["azure-gpt-3.5"]
duration: None
cache: True # set cache responses to True
cache_params:
type: "redis-semantic"
similarity_threshold: 0.8
redis_semantic_cache_use_async: True
# cache: True
# setting callback class
# callbacks: custom_callbacks.proxy_handler_instance # sets litellm.callbacks = [proxy_handler_instance]