litellm-mirror/litellm/tests/test_proxy.py
2023-10-20 09:21:59 -07:00

38 lines
No EOL
1.4 KiB
Python

#### What this tests ####
# This tests the OpenAI-proxy server
import sys, os
import traceback
sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the system path
from dotenv import load_dotenv
load_dotenv()
import unittest
from unittest.mock import patch
from click.testing import CliRunner
import pytest
import litellm
from litellm.proxy.llm import litellm_completion
from litellm.proxy.proxy_server import initialize
def test_azure_call():
try:
data = {
"model": "azure/chatgpt-v-2",
"messages": [{"role": "user", "content": "Hey!"}]
}
result = litellm_completion(data=data, user_api_base=os.getenv("AZURE_API_BASE"), type="chat_completion", user_temperature=None, user_max_tokens=None, user_model=None, user_headers=None, user_debug=False, model_router=None)
return result
except Exception as e:
pytest.fail(f"An error occurred: {e}")
# test_azure_call()
## test debug
def test_debug():
try:
initialize(model=None, alias=None, api_base=None, debug=True, temperature=None, max_tokens=None, max_budget=None, telemetry=None, drop_params=None, add_function_to_prompt=None, headers=None, save=None, api_version=None)
assert litellm.set_verbose == True
except Exception as e:
pytest.fail(f"An error occurred: {e}")
# test_debug()
## test logs