litellm-mirror/litellm/tests/test_secret_detect_hook.py
2024-06-25 17:27:02 -07:00

216 lines
6.4 KiB
Python

# What is this?
## This tests the llm guard integration
import asyncio
import os
import random
# What is this?
## Unit test for presidio pii masking
import sys
import time
import traceback
from datetime import datetime
from dotenv import load_dotenv
load_dotenv()
import os
sys.path.insert(
0, os.path.abspath("../..")
) # Adds the parent directory to the system path
import pytest
import litellm
from litellm import Router, mock_completion
from litellm.caching import DualCache
from litellm.proxy._types import UserAPIKeyAuth
from litellm.proxy.enterprise.enterprise_hooks.secret_detection import (
_ENTERPRISE_SecretDetection,
)
from litellm.proxy.utils import ProxyLogging, hash_token
### UNIT TESTS FOR OpenAI Moderation ###
@pytest.mark.asyncio
async def test_basic_secret_detection_chat():
"""
Tests to see if secret detection hook will mask api keys
It should mask the following API_KEY = 'sk_1234567890abcdef' and OPENAI_API_KEY = 'sk_1234567890abcdef'
"""
secret_instance = _ENTERPRISE_SecretDetection()
_api_key = "sk-12345"
_api_key = hash_token("sk-12345")
user_api_key_dict = UserAPIKeyAuth(api_key=_api_key)
local_cache = DualCache()
from litellm.proxy.proxy_server import llm_router
test_data = {
"messages": [
{
"role": "user",
"content": "Hey, how's it going, API_KEY = 'sk_1234567890abcdef'",
},
{
"role": "assistant",
"content": "Hello! I'm doing well. How can I assist you today?",
},
{
"role": "user",
"content": "this is my OPENAI_API_KEY = 'sk_1234567890abcdef'",
},
{"role": "user", "content": "i think it is +1 412-555-5555"},
],
"model": "gpt-3.5-turbo",
}
await secret_instance.async_pre_call_hook(
cache=local_cache,
data=test_data,
user_api_key_dict=user_api_key_dict,
call_type="completion",
)
print(
"test data after running pre_call_hook: Expect all API Keys to be masked",
test_data,
)
assert test_data == {
"messages": [
{"role": "user", "content": "Hey, how's it going, API_KEY = '[REDACTED]'"},
{
"role": "assistant",
"content": "Hello! I'm doing well. How can I assist you today?",
},
{"role": "user", "content": "this is my OPENAI_API_KEY = '[REDACTED]'"},
{"role": "user", "content": "i think it is +1 412-555-5555"},
],
"model": "gpt-3.5-turbo",
}, "Expect all API Keys to be masked"
@pytest.mark.asyncio
async def test_basic_secret_detection_text_completion():
"""
Tests to see if secret detection hook will mask api keys
It should mask the following API_KEY = 'sk_1234567890abcdef' and OPENAI_API_KEY = 'sk_1234567890abcdef'
"""
secret_instance = _ENTERPRISE_SecretDetection()
_api_key = "sk-12345"
_api_key = hash_token("sk-12345")
user_api_key_dict = UserAPIKeyAuth(api_key=_api_key)
local_cache = DualCache()
from litellm.proxy.proxy_server import llm_router
test_data = {
"prompt": "Hey, how's it going, API_KEY = 'sk_1234567890abcdef', my OPENAI_API_KEY = 'sk_1234567890abcdef' and i want to know what is the weather",
"model": "gpt-3.5-turbo",
}
await secret_instance.async_pre_call_hook(
cache=local_cache,
data=test_data,
user_api_key_dict=user_api_key_dict,
call_type="completion",
)
test_data == {
"prompt": "Hey, how's it going, API_KEY = '[REDACTED]', my OPENAI_API_KEY = '[REDACTED]' and i want to know what is the weather",
"model": "gpt-3.5-turbo",
}
print(
"test data after running pre_call_hook: Expect all API Keys to be masked",
test_data,
)
@pytest.mark.asyncio
async def test_basic_secret_detection_embeddings():
"""
Tests to see if secret detection hook will mask api keys
It should mask the following API_KEY = 'sk_1234567890abcdef' and OPENAI_API_KEY = 'sk_1234567890abcdef'
"""
secret_instance = _ENTERPRISE_SecretDetection()
_api_key = "sk-12345"
_api_key = hash_token("sk-12345")
user_api_key_dict = UserAPIKeyAuth(api_key=_api_key)
local_cache = DualCache()
from litellm.proxy.proxy_server import llm_router
test_data = {
"input": "Hey, how's it going, API_KEY = 'sk_1234567890abcdef', my OPENAI_API_KEY = 'sk_1234567890abcdef' and i want to know what is the weather",
"model": "gpt-3.5-turbo",
}
await secret_instance.async_pre_call_hook(
cache=local_cache,
data=test_data,
user_api_key_dict=user_api_key_dict,
call_type="embedding",
)
assert test_data == {
"input": "Hey, how's it going, API_KEY = '[REDACTED]', my OPENAI_API_KEY = '[REDACTED]' and i want to know what is the weather",
"model": "gpt-3.5-turbo",
}
print(
"test data after running pre_call_hook: Expect all API Keys to be masked",
test_data,
)
@pytest.mark.asyncio
async def test_basic_secret_detection_embeddings_list():
"""
Tests to see if secret detection hook will mask api keys
It should mask the following API_KEY = 'sk_1234567890abcdef' and OPENAI_API_KEY = 'sk_1234567890abcdef'
"""
secret_instance = _ENTERPRISE_SecretDetection()
_api_key = "sk-12345"
_api_key = hash_token("sk-12345")
user_api_key_dict = UserAPIKeyAuth(api_key=_api_key)
local_cache = DualCache()
from litellm.proxy.proxy_server import llm_router
test_data = {
"input": [
"hey",
"how's it going, API_KEY = 'sk_1234567890abcdef'",
"my OPENAI_API_KEY = 'sk_1234567890abcdef' and i want to know what is the weather",
],
"model": "gpt-3.5-turbo",
}
await secret_instance.async_pre_call_hook(
cache=local_cache,
data=test_data,
user_api_key_dict=user_api_key_dict,
call_type="embedding",
)
print(
"test data after running pre_call_hook: Expect all API Keys to be masked",
test_data,
)
assert test_data == {
"input": [
"hey",
"how's it going, API_KEY = '[REDACTED]'",
"my OPENAI_API_KEY = '[REDACTED]' and i want to know what is the weather",
],
"model": "gpt-3.5-turbo",
}