forked from phoenix/litellm-mirror
fix litellm client
This commit is contained in:
parent
8f5e2d0013
commit
e452ceb21e
4 changed files with 8 additions and 3 deletions
Binary file not shown.
Binary file not shown.
|
@ -30,7 +30,7 @@ sys.path.insert(
|
||||||
0, os.path.abspath("../..")
|
0, os.path.abspath("../..")
|
||||||
) # Adds the parent directory to the system path
|
) # Adds the parent directory to the system path
|
||||||
import litellm
|
import litellm
|
||||||
from litellm import completion_with_split_tests
|
from litellm import completion
|
||||||
from openai.error import AuthenticationError
|
from openai.error import AuthenticationError
|
||||||
litellm.set_verbose = True
|
litellm.set_verbose = True
|
||||||
|
|
||||||
|
@ -69,8 +69,8 @@ messages = [{"content": user_message, "role": "user"}]
|
||||||
# # Redirect stdout
|
# # Redirect stdout
|
||||||
# old_stdout = sys.stdout
|
# old_stdout = sys.stdout
|
||||||
# sys.stdout = new_stdout = io.StringIO()
|
# sys.stdout = new_stdout = io.StringIO()
|
||||||
|
# litellm.token = "a67abbaf-35b8-4649-8647-68c5fe8d37fb" # generate one here - https://www.uuidgenerator.net/version4
|
||||||
# response = completion_with_split_tests(models=split_per_model, messages=messages, use_client=True, id="6d383c99-488d-481d-aa1b-1f94935cec44")
|
# response = completion(model="gpt-3.5-turbo", messages=messages, use_client=True)
|
||||||
|
|
||||||
# # Restore stdout
|
# # Restore stdout
|
||||||
# sys.stdout = old_stdout
|
# sys.stdout = old_stdout
|
||||||
|
@ -83,6 +83,7 @@ messages = [{"content": user_message, "role": "user"}]
|
||||||
# if "LiteDebugger: Success/Failure Call Logging" not in output:
|
# if "LiteDebugger: Success/Failure Call Logging" not in output:
|
||||||
# raise Exception("LiteLLMDebugger: success/failure call not logged!")
|
# raise Exception("LiteLLMDebugger: success/failure call not logged!")
|
||||||
# except Exception as e:
|
# except Exception as e:
|
||||||
|
# print(output)
|
||||||
# pytest.fail(f"Error occurred: {e}")
|
# pytest.fail(f"Error occurred: {e}")
|
||||||
|
|
||||||
# # Test 3: On streaming completion call - setting client to true
|
# # Test 3: On streaming completion call - setting client to true
|
||||||
|
|
|
@ -1051,6 +1051,10 @@ def set_callbacks(callback_list, function_id=None):
|
||||||
print_verbose(f"instantiating lite_debugger")
|
print_verbose(f"instantiating lite_debugger")
|
||||||
if function_id:
|
if function_id:
|
||||||
liteDebuggerClient = LiteDebugger(email=function_id)
|
liteDebuggerClient = LiteDebugger(email=function_id)
|
||||||
|
elif litellm.token:
|
||||||
|
liteDebuggerClient = LiteDebugger(email=litellm.token)
|
||||||
|
elif litellm.email:
|
||||||
|
liteDebuggerClient = LiteDebugger(email=litellm.email)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue