mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
fix(test_text_completion.py): fix test
This commit is contained in:
parent
ff12e023ae
commit
fd6ccfca7d
3 changed files with 14 additions and 8 deletions
|
@ -2267,11 +2267,16 @@ def client(original_function):
|
|||
and kwargs["complete_response"] == True
|
||||
):
|
||||
chunks = []
|
||||
for idx, chunk in enumerate(result):
|
||||
async for chunk in result:
|
||||
chunks.append(chunk)
|
||||
return litellm.stream_chunk_builder(
|
||||
chunks, messages=kwargs.get("messages", None)
|
||||
)
|
||||
if call_type == CallTypes.acompletion.value:
|
||||
return litellm.stream_chunk_builder(
|
||||
chunks, messages=kwargs.get("messages")
|
||||
)
|
||||
elif call_type == CallTypes.atext_completion.value:
|
||||
return litellm.stream_chunk_builder(
|
||||
chunks, messages=[{"role": "user", "content": kwargs.get("prompt")}]
|
||||
)
|
||||
else:
|
||||
return result
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue