(test) async cloudflare

This commit is contained in:
ishaan-jaff 2023-12-29 11:49:07 +05:30
parent dde6bc4fb6
commit a999e80b46
2 changed files with 25 additions and 7 deletions

View file

@ -130,6 +130,30 @@ def test_async_anyscale_response():
# test_async_anyscale_response() # test_async_anyscale_response()
def test_async_completion_cloudflare():
try:
litellm.set_verbose = True
async def test():
response = await litellm.acompletion(
model="cloudflare/@cf/meta/llama-2-7b-chat-int8",
messages=[{"content": "what llm are you", "role": "user"}],
max_tokens=50,
)
print(response)
return response
response = asyncio.run(test())
text_response = response["choices"][0]["message"]["content"]
assert len(text_response) > 5 # more than 5 chars in response
except Exception as e:
pytest.fail(f"Error occurred: {e}")
test_async_completion_cloudflare()
def test_get_response_streaming(): def test_get_response_streaming():
import asyncio import asyncio

View file

@ -2009,6 +2009,7 @@ def test_completion_cloudflare():
response = completion( response = completion(
model="cloudflare/@cf/meta/llama-2-7b-chat-int8", model="cloudflare/@cf/meta/llama-2-7b-chat-int8",
messages=[{"content": "what llm are you", "role": "user"}], messages=[{"content": "what llm are you", "role": "user"}],
max_tokens=15,
) )
print(response) print(response)
@ -2018,13 +2019,6 @@ def test_completion_cloudflare():
test_completion_cloudflare() test_completion_cloudflare()
# async def get_response(generator):
# async for elem in generator:
# print(elem)
# return
# test_completion_together_ai_stream()
def test_moderation(): def test_moderation():
import openai import openai