forked from phoenix/litellm-mirror
stream response (#5516)
This commit is contained in:
parent
aefca63a8c
commit
258b96211c
1 changed files with 9 additions and 0 deletions
|
@ -868,6 +868,15 @@ class Logging:
|
||||||
model = self.model
|
model = self.model
|
||||||
messages = self.model_call_details["input"]
|
messages = self.model_call_details["input"]
|
||||||
kwargs = self.model_call_details
|
kwargs = self.model_call_details
|
||||||
|
|
||||||
|
# this only logs streaming once, complete_streaming_response exists i.e when stream ends
|
||||||
|
if self.stream:
|
||||||
|
if "complete_streaming_response" not in kwargs:
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print_verbose("reaches helicone for streaming logging!")
|
||||||
|
result = kwargs["complete_streaming_response"]
|
||||||
|
|
||||||
heliconeLogger.log_success(
|
heliconeLogger.log_success(
|
||||||
model=model,
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue