litellm/tests/local_testing/test_wandb.py

72 lines
1.7 KiB
Python

import sys
import os
import io, asyncio
# import logging
# logging.basicConfig(level=logging.DEBUG)
sys.path.insert(0, os.path.abspath("../.."))
from litellm import completion
import litellm
litellm.num_retries = 3
litellm.success_callback = ["wandb"]
import time
import pytest
def test_wandb_logging_async():
try:
litellm.set_verbose = False
async def _test_langfuse():
from litellm import Router
model_list = [
{ # list of model deployments
"model_name": "gpt-3.5-turbo",
"litellm_params": { # params for litellm completion/embedding call
"model": "gpt-3.5-turbo",
"api_key": os.getenv("OPENAI_API_KEY"),
},
}
]
router = Router(model_list=model_list)
# openai.ChatCompletion.create replacement
response = await router.acompletion(
model="gpt-3.5-turbo",
messages=[
{"role": "user", "content": "this is a test with litellm router ?"}
],
)
print(response)
response = asyncio.run(_test_langfuse())
print(f"response: {response}")
except litellm.Timeout as e:
pass
except Exception as e:
pass
test_wandb_logging_async()
def test_wandb_logging():
try:
response = completion(
model="claude-instant-1.2",
messages=[{"role": "user", "content": "Hi 👋 - i'm claude"}],
max_tokens=10,
temperature=0.2,
)
print(response)
except litellm.Timeout as e:
pass
except Exception as e:
print(e)
# test_wandb_logging()