import sys import os import io, asyncio # import logging # logging.basicConfig(level=logging.DEBUG) sys.path.insert(0, os.path.abspath('../..')) from litellm import completion import litellm litellm.num_retries = 3 litellm.success_callback = ["wandb"] import time import pytest def test_wandb_logging_async(): try: litellm.set_verbose = False async def _test_langfuse(): from litellm import Router model_list = [{ # list of model deployments "model_name": "gpt-3.5-turbo", "litellm_params": { # params for litellm completion/embedding call "model": "gpt-3.5-turbo", "api_key": os.getenv("OPENAI_API_KEY"), } }] router = Router(model_list=model_list) # openai.ChatCompletion.create replacement response = await router.acompletion(model="gpt-3.5-turbo", messages=[{"role": "user", "content": "this is a test with litellm router ?"}]) print(response) response = asyncio.run(_test_langfuse()) print(f"response: {response}") except litellm.Timeout as e: pass except Exception as e: pass test_wandb_logging_async() def test_wandb_logging(): try: response = completion(model="claude-instant-1.2", messages=[{ "role": "user", "content": "Hi 👋 - i'm claude" }], max_tokens=10, temperature=0.2 ) print(response) except litellm.Timeout as e: pass except Exception as e: print(e) # test_wandb_logging()