mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
add metadata logging to promptlayer
This commit is contained in:
parent
cf97a9441a
commit
b76372cee0
8 changed files with 100 additions and 20 deletions
|
@ -177,7 +177,7 @@ class CallTypes(Enum):
|
|||
|
||||
# Logging function -> log the exact model details + what's being sent | Non-Blocking
|
||||
class Logging:
|
||||
global supabaseClient, liteDebuggerClient
|
||||
global supabaseClient, liteDebuggerClient, promptLayerLogger
|
||||
|
||||
def __init__(self, model, messages, stream, call_type, start_time, litellm_call_id, function_id):
|
||||
if call_type not in [item.value for item in CallTypes]:
|
||||
|
@ -395,6 +395,15 @@ class Logging:
|
|||
self.litellm_params["stream_response"][litellm_call_id] = new_model_response
|
||||
#print("adding to cache for", litellm_call_id)
|
||||
litellm.cache.add_cache(self.litellm_params["stream_response"][litellm_call_id], **self.model_call_details)
|
||||
if callback == "promptlayer":
|
||||
print_verbose("reaches promptlayer for logging!")
|
||||
promptLayerLogger.log_event(
|
||||
kwargs=self.model_call_details,
|
||||
response_obj=result,
|
||||
start_time=start_time,
|
||||
end_time=end_time,
|
||||
print_verbose=print_verbose,
|
||||
)
|
||||
|
||||
except Exception as e:
|
||||
print_verbose(
|
||||
|
@ -827,7 +836,8 @@ def get_litellm_params(
|
|||
api_base=None,
|
||||
litellm_call_id=None,
|
||||
model_alias_map=None,
|
||||
completion_call_id=None
|
||||
completion_call_id=None,
|
||||
metadata=None
|
||||
):
|
||||
litellm_params = {
|
||||
"return_async": return_async,
|
||||
|
@ -840,6 +850,7 @@ def get_litellm_params(
|
|||
"litellm_call_id": litellm_call_id,
|
||||
"model_alias_map": model_alias_map,
|
||||
"completion_call_id": completion_call_id,
|
||||
"metadata": metadata,
|
||||
"stream_response": {} # litellm_call_id: ModelResponse Dict
|
||||
}
|
||||
|
||||
|
@ -1630,15 +1641,6 @@ def handle_success(args, kwargs, result, start_time, end_time):
|
|||
run_id=kwargs["litellm_call_id"],
|
||||
print_verbose=print_verbose,
|
||||
)
|
||||
elif callback == "promptlayer":
|
||||
print_verbose("reaches promptlayer for logging!")
|
||||
promptLayerLogger.log_event(
|
||||
kwargs=kwargs,
|
||||
response_obj=result,
|
||||
start_time=start_time,
|
||||
end_time=end_time,
|
||||
print_verbose=print_verbose,
|
||||
)
|
||||
elif callback == "langfuse":
|
||||
print_verbose("reaches langfuse for logging!")
|
||||
langFuseLogger.log_event(
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue