forked from phoenix/litellm-mirror
(feat) working semantic cache on proxy
This commit is contained in:
parent
a3b1e3bc84
commit
70a895329e
1 changed files with 6 additions and 1 deletions
|
@ -73,6 +73,11 @@ litellm_settings:
|
|||
max_budget: 1.5000
|
||||
models: ["azure-gpt-3.5"]
|
||||
duration: None
|
||||
cache: True # set cache responses to True
|
||||
cache_params:
|
||||
type: "redis-semantic"
|
||||
similarity_threshold: 0.8
|
||||
redis_semantic_cache_use_async: True
|
||||
# cache: True
|
||||
# setting callback class
|
||||
# callbacks: custom_callbacks.proxy_handler_instance # sets litellm.callbacks = [proxy_handler_instance]
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue