simplifying calling litellm debugger

This commit is contained in:
Krrish Dholakia 2023-08-21 11:37:07 -07:00
parent 7da8dfc820
commit 3286a708fe
7 changed files with 12 additions and 17 deletions

View file

@ -16,14 +16,13 @@ See our live dashboard 👉 [admin.litellm.ai](https://admin.litellm.ai/)
By default, your dashboard is viewable at `admin.litellm.ai/<your_email>`.
```
import litellm, os
## Set your email
os.environ["LITELLM_EMAIL"] = "your_user_email"
## LOG ON ALL 3 EVENTS
litellm.input_callback = ["lite_debugger"]
litellm.success_callback = ["lite_debugger"]
litellm.failure_callback = ["lite_debugger"]
## Set debugger to true
litellm.debugger = True
```
## Example Usage
@ -36,12 +35,8 @@ By default, your dashboard is viewable at `admin.litellm.ai/<your_email>`.
## Set ENV variable
os.environ["LITELLM_EMAIL"] = "your_email"
## LOG ON ALL 3 EVENTS
litellm.input_callback = ["lite_debugger"]
litellm.success_callback = ["lite_debugger"]
litellm.failure_callback = ["lite_debugger"]
litellm.set_verbose = True
## Set debugger to true
litellm.debugger = True
user_message = "Hello, how are you?"
messages = [{ "content": user_message,"role": "user"}]
@ -52,6 +47,5 @@ By default, your dashboard is viewable at `admin.litellm.ai/<your_email>`.
# bad request call
response = completion(model="chatgpt-test", messages=[{"role": "user", "content": "Hi 👋 - i'm a bad request"}])
```

View file

@ -21,6 +21,7 @@ hugging_api_token: Optional[str] = None
togetherai_api_key: Optional[str] = None
caching = False
caching_with_models = False # if you want the caching key to be model + prompt
debugger = False
model_cost = {
"gpt-3.5-turbo": {
"max_tokens": 4000,

View file

@ -9,11 +9,7 @@
# import litellm
# from litellm import embedding, completion
# litellm.input_callback = ["lite_debugger"]
# litellm.success_callback = ["lite_debugger"]
# litellm.failure_callback = ["lite_debugger"]
# litellm.set_verbose = True
# litellm.debugger = True
# user_message = "Hello, how are you?"
# messages = [{ "content": user_message,"role": "user"}]

View file

@ -286,6 +286,10 @@ def client(original_function):
): # just run once to check if user wants to send their data anywhere - PostHog/Sentry/Slack/etc.
try:
global callback_list, add_breadcrumb, user_logger_fn
if litellm.debugger: # add to input, success and failure callbacks if user sets debugging to true
litellm.input_callback.append("lite_debugger")
litellm.success_callback.append("lite_debugger")
litellm.failure_callback.append("lite_debugger")
if (
len(litellm.input_callback) > 0 or len(litellm.success_callback) > 0 or len(litellm.failure_callback) > 0
) and len(callback_list) == 0: