mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
Update test
This commit is contained in:
parent
33a047315e
commit
abb8e17d4b
1 changed files with 15 additions and 11 deletions
|
@ -460,23 +460,27 @@ def test_completion_azure_stream_content_filter_no_delta():
|
|||
|
||||
response = litellm.CustomStreamWrapper(
|
||||
completion_stream=completion_stream,
|
||||
model="azure/gpt-4o",
|
||||
model="gpt-4-0613",
|
||||
custom_llm_provider="cached_response",
|
||||
messages=[{"role": "user", "content": "Hey"}],
|
||||
stream=True,
|
||||
call_type="completion",
|
||||
start_time=time.time(),
|
||||
litellm_call_id="12345",
|
||||
function_id="1245",
|
||||
),
|
||||
logging_obj=litellm.Logging(
|
||||
model="gpt-4-0613",
|
||||
messages=[{"role": "user", "content": "Hey"}],
|
||||
stream=True,
|
||||
call_type="completion",
|
||||
start_time=time.time(),
|
||||
litellm_call_id="12345",
|
||||
function_id="1245",
|
||||
),
|
||||
)
|
||||
|
||||
for idx, chunk in enumerate(response):
|
||||
complete_response = ""
|
||||
for idx, chunk in enumerate(response):
|
||||
# print
|
||||
chunk, finished = streaming_format_tests(idx, chunk)
|
||||
complete_response += chunk
|
||||
if finished:
|
||||
delta = chunk.choices[0].delta
|
||||
content = delta.content if delta else None
|
||||
complete_response += content or ""
|
||||
if chunk.choices[0].finish_reason is not None:
|
||||
break
|
||||
assert len(complete_response) > 0
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue