From 80fa61cc0ebe8f9a7c706c88b4693a65d2e0171d Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Mon, 28 Aug 2023 15:36:57 -0700 Subject: [PATCH] bug fixes to logging --- litellm/__pycache__/utils.cpython-311.pyc | Bin 70266 -> 70364 bytes .../tests/test_litedebugger_integration.py | 2 +- litellm/utils.py | 2 +- 3 files changed, 2 insertions(+), 2 deletions(-) diff --git a/litellm/__pycache__/utils.cpython-311.pyc b/litellm/__pycache__/utils.cpython-311.pyc index a752c8f58bae12599ddd075f2a20e2c41c39ce06..b2edcbec9b212b45afc6408c048ea239439141c9 100644 GIT binary patch delta 473 zcmeyhgyqgs7QW@Yyj%=GP@wQOMOc0#U&mc879f`yh(E8Iyx^`PUkNKn9s*JrCnxR@ z)2U&pVTcDQ0fQQ*5)K%ffgua7yo{kptb}(m<54MBpmKJYk`#tuh8jk65ltq)DuJB* z^z_WUbcOt+tkmQZD+NuCB0HeAqO%~vYqI&hTAgGdvw`7;s8k0_57!N4vksOk913?i zcrWnjEszEhKy;Bq{|blx1rGhmC+{s`{4+W4{!zyNn@t{EWn^UD%=Ji#iIHow?h`g9 zM$yT-Ph>U+JbS~)D8JeEMKdF#`sRZ#pE5D(ZeISzh>_86^R2h6EKFGflQS2{PyY5X zk5OoH>ZfR?qM4KTeu`$ZVPjw@cHJ!ZS%yjS3WvmX4y8*RN(<~ZXkJmaxyWIAg~Rp& zhwbEmue~CitmYpWPzkZkAHQ}mF(z%U`Z1N2v1GHrzXi;U)3$G9WXxn?oV#6`jqxV8 ZK^`Nc#RmrLwbWn1Ngn{@gZs!Cggx8kQP{cqWh<5U63A zys$-#t&E{atb}{A;5}JgHi!fRLkdGMLk%NPF^Wi0HzNasCVP<`P+8I0$p!anMSXzG z28J7=QXMQkTsM@>I#{l7C`^{Tzl8Dk=GxCPOk&qLBrb7C%#d5bc}3FbB8Tx64&w_P#*;UF?G@o* zHUGeXN{DO@{?@_7n7H}QkEyJT#hZ)%EnsGxx?PxwF_VRH&h}b1#+%#*xr~e!9~iKc RAHm{ZkY$smw+S(Z0RXB1c*g($ diff --git a/litellm/tests/test_litedebugger_integration.py b/litellm/tests/test_litedebugger_integration.py index 3eca24361..90e7e5bdc 100644 --- a/litellm/tests/test_litedebugger_integration.py +++ b/litellm/tests/test_litedebugger_integration.py @@ -22,7 +22,7 @@ response = completion(model="claude-instant-1", messages=[{"role": "user", "cont # print(f"response: {response}") # # Test 2: On embedding call -# response = embedding(model="text-embedding-ada-002", input=["sample text"]) +response = embedding(model="text-embedding-ada-002", input=["sample text"]) # print(f"response: {response}") # # Test 3: On streaming completion call diff --git a/litellm/utils.py b/litellm/utils.py index 3ef075a18..0cc826bfe 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -1627,7 +1627,7 @@ class CustomStreamWrapper: chunk = next(self.completion_stream) completion_obj['content'] = chunk['choices']['delta'] # LOGGING - self.logging_obj(completion_obj["content"]) + self.logging_obj.post_call(completion_obj["content"]) # return this for all models return {"choices": [{"delta": completion_obj}]}