litellm-mirror/litellm/tests/test_budget_manager.py
2023-09-11 13:46:27 -07:00

70 lines
No EOL
2.3 KiB
Python

#### What this tests ####
# This tests calling batch_completions by running 100 messages together
import sys, os, json
import traceback
import pytest
sys.path.insert(
0, os.path.abspath("../..")
) # Adds the parent directory to the system path
import litellm
from litellm import BudgetManager, completion
budget_manager = BudgetManager(project_name="test_project")
## Scenario 1: User budget enough to make call
def test_user_budget_enough():
try:
user = "1234"
# create a budget for a user
budget_manager.create_budget(total_budget=10, user=user)
# check if a given call can be made
data = {
"model": "gpt-3.5-turbo",
"messages": [{"role": "user", "content": "Hey, how's it going?"}]
}
if budget_manager.get_current_cost(user=user) <= budget_manager.get_total_budget(user):
response = completion(**data)
print(budget_manager.update_cost(completion_obj=response, user=user))
else:
response = "Sorry - no budget!"
print(f"response: {response}")
except:
pytest.fail(f"An error occurred")
## Scenario 2: User budget not enough to make call
def test_user_budget_not_enough():
try:
user = "12345"
# create a budget for a user
budget_manager.create_budget(total_budget=0, user=user)
# check if a given call can be made
data = {
"model": "gpt-3.5-turbo",
"messages": [{"role": "user", "content": "Hey, how's it going?"}]
}
model = data["model"]
messages = data["messages"]
if budget_manager.get_current_cost(user=user) < budget_manager.get_total_budget(user=user):
response = completion(**data)
print(budget_manager.update_cost(completion_obj=response, user=user))
else:
response = "Sorry - no budget!"
print(f"response: {response}")
except:
pytest.fail(f"An error occurred")
## Scenario 3: Saving budget to client
def test_save_user_budget():
try:
response = budget_manager.save_data()
if response["status"] == "error":
raise Exception(f"An error occurred - {json.dumps(response)}")
print(response)
except:
pytest.fail(f"An error occurred")