# #### What this tests #### # # This tests using caching w/ litellm which requires SSL=True # import sys, os # import time # import traceback # from dotenv import load_dotenv # load_dotenv() # import os # sys.path.insert( # 0, os.path.abspath("../..") # ) # Adds the parent directory to the system path # import pytest # import litellm # from litellm import embedding, completion # from litellm.caching import Cache # messages = [{"role": "user", "content": f"who is ishaan {time.time()}"}] # @pytest.mark.skip(reason="local proxy test") # def test_caching_v2(): # test in memory cache # try: # response1 = completion(model="openai/gpt-3.5-turbo", messages=messages, api_base="http://0.0.0.0:8000") # response2 = completion(model="openai/gpt-3.5-turbo", messages=messages, api_base="http://0.0.0.0:8000") # print(f"response1: {response1}") # print(f"response2: {response2}") # litellm.cache = None # disable cache # if response2['choices'][0]['message']['content'] != response1['choices'][0]['message']['content']: # print(f"response1: {response1}") # print(f"response2: {response2}") # raise Exception() # except Exception as e: # print(f"error occurred: {traceback.format_exc()}") # pytest.fail(f"Error occurred: {e}") # test_caching_v2()