From 0d44f5e441364103de452cd85d7355d70f6beba6 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Mon, 4 Dec 2023 09:23:55 -0800 Subject: [PATCH] (feat) proxy:custom_logger --- litellm/proxy/custom_logger.py | 33 ++++++++++++++++++++++----------- 1 file changed, 22 insertions(+), 11 deletions(-) diff --git a/litellm/proxy/custom_logger.py b/litellm/proxy/custom_logger.py index d30722bd9f..5442837235 100644 --- a/litellm/proxy/custom_logger.py +++ b/litellm/proxy/custom_logger.py @@ -5,24 +5,35 @@ class MyCustomHandler(CustomLogger): print(f"Pre-API Call") def log_post_api_call(self, kwargs, response_obj, start_time, end_time): - # log: key, user, model, prompt, response, tokens, cost print(f"Post-API Call") - print("\n kwargs\n") - print(kwargs) - model = kwargs["model"] - messages = kwargs["messages"] - cost = litellm.completion_cost(completion_response=response_obj) - # tokens used in response - usage = response_obj.usage - print(usage) - - def log_stream_event(self, kwargs, response_obj, start_time, end_time): print(f"On Stream") def log_success_event(self, kwargs, response_obj, start_time, end_time): print(f"On Success") + # log: key, user, model, prompt, response, tokens, cost + print("\n kwargs\n") + print(kwargs) + model = kwargs["model"] + messages = kwargs["messages"] + cost = litellm.completion_cost(completion_response=response_obj) + response = response_obj + # tokens used in response + usage = response_obj["usage"] + + print( + f""" + Model: {model}, + Messages: {messages}, + Usage: {usage}, + Cost: {cost}, + Response: {response} + """ + ) + + print(usage) + def log_failure_event(self, kwargs, response_obj, start_time, end_time): print(f"On Failure")