mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
25 lines
No EOL
785 B
Python
25 lines
No EOL
785 B
Python
# #### What this tests ####
|
|
# # This tests calling batch_completions by running 100 messages together
|
|
|
|
# import sys, os
|
|
# import traceback
|
|
# import pytest
|
|
# sys.path.insert(
|
|
# 0, os.path.abspath("../..")
|
|
# ) # Adds the parent directory to the system path
|
|
# from openai.error import Timeout
|
|
# import litellm
|
|
# from litellm import batch_completion
|
|
# litellm.set_verbose=True
|
|
|
|
# def test_batch_completions():
|
|
# messages = [[{"role": "user", "content": "Hey, how's it going"}] for _ in range(5)]
|
|
# model = "gpt-3.5-turbo"
|
|
# try:
|
|
# result = batch_completion(model=model, messages=messages)
|
|
# print(result)
|
|
# print(len(result))
|
|
# except Timeout as e:
|
|
# pass
|
|
# except Exception as e:
|
|
# pytest.fail(f"An error occurred: {e}") |