import sys import os import io, asyncio # import logging # logging.basicConfig(level=logging.DEBUG) sys.path.insert(0, os.path.abspath("../..")) from litellm import completion import litellm litellm.num_retries = 3 litellm.success_callback = ["wandb"] import time import pytest def test_wandb_logging_async(): try: litellm.set_verbose = False async def _test_langfuse(): from litellm import Router model_list = [ { # list of model deployments "model_name": "gpt-3.5-turbo", "litellm_params": { # params for litellm completion/embedding call "model": "gpt-3.5-turbo", "api_key": os.getenv("OPENAI_API_KEY"), }, } ] router = Router(model_list=model_list) # openai.ChatCompletion.create replacement response = await router.acompletion( model="gpt-3.5-turbo", messages=[ {"role": "user", "content": "this is a test with litellm router ?"} ], ) print(response) response = asyncio.run(_test_langfuse()) print(f"response: {response}") except litellm.Timeout as e: pass except Exception as e: pass test_wandb_logging_async() def test_wandb_logging(): try: response = completion( model="claude-3-5-haiku-20241022", messages=[{"role": "user", "content": "Hi 👋 - i'm claude"}], max_tokens=10, temperature=0.2, ) print(response) except litellm.Timeout as e: pass except Exception as e: print(e) # test_wandb_logging()