#### What this tests #### # This tests if logging to the helicone integration actually works import sys, os import traceback import pytest sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the system path import litellm from litellm import embedding, completion litellm.success_callback = ["helicone"] litellm.set_verbose = True user_message = "Hello, how are you?" messages = [{ "content": user_message,"role": "user"}] #openai call response = completion(model="gpt-3.5-turbo", messages=[{"role": "user", "content": "Hi 👋 - i'm openai"}]) #cohere call response = completion(model="command-nightly", messages=[{"role": "user", "content": "Hi 👋 - i'm cohere"}])