mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix(router.py): skip setting model_group response headers for now
current implementation increases redis cache calls by 3x
This commit is contained in:
parent
5fbcdd8b11
commit
81d6c5e5a5
4 changed files with 43 additions and 89 deletions
|
@ -1860,6 +1860,7 @@ class DualCache(BaseCache):
|
|||
|
||||
Returns - int - the incremented value
|
||||
"""
|
||||
traceback.print_stack()
|
||||
try:
|
||||
result: int = value
|
||||
if self.in_memory_cache is not None:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue