fixes to caching, testing caching

This commit is contained in:
ishaan-jaff 2023-08-17 10:41:14 -07:00
parent f9e9108482
commit 701dd83853
4 changed files with 57 additions and 23 deletions

View file

@ -8,8 +8,8 @@ import litellm
from litellm import embedding, completion
from infisical import InfisicalClient
# litellm.set_verbose = True
litellm.secret_manager_client = InfisicalClient(token=os.environ["INFISICAL_TOKEN"])
# # litellm.set_verbose = True
# litellm.secret_manager_client = InfisicalClient(token=os.environ["INFISICAL_TOKEN"])
def test_openai_embedding():
try: