fix caching

This commit is contained in:
ishaan-jaff 2023-08-28 14:53:41 -07:00
parent 1799f1bfe9
commit 83aa304c34
3 changed files with 2 additions and 2 deletions

Binary file not shown.

View file

@ -59,12 +59,12 @@ def test_caching_with_models():
def test_gpt_cache(): def test_gpt_cache():
# INIT GPT Cache # # INIT GPT Cache #
from gptcache import cache from gptcache import cache
from litellm.cache import completion from litellm.gpt_cache import completion
cache.init() cache.init()
cache.set_openai_key() cache.set_openai_key()
messages = [{"role": "user", "content": "what is litellm YC 22?"}] messages = [{"role": "user", "content": "what is litellm YC paul graham, partner?"}]
response2 = completion(model="gpt-3.5-turbo", messages=messages) response2 = completion(model="gpt-3.5-turbo", messages=messages)
response3 = completion(model="command-nightly", messages=messages) response3 = completion(model="command-nightly", messages=messages)
print(f"response2: {response2}") print(f"response2: {response2}")