mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 07:32:36 +00:00
add nvidia provider for inference tests
This commit is contained in:
parent
1e18791fff
commit
5ab3773577
2 changed files with 19 additions and 1 deletions
|
|
@ -361,7 +361,10 @@ class TestInference:
|
|||
for chunk in grouped[ChatCompletionResponseEventType.progress]
|
||||
)
|
||||
first = grouped[ChatCompletionResponseEventType.progress][0]
|
||||
assert first.event.delta.parse_status == ToolCallParseStatus.started
|
||||
if not isinstance(
|
||||
first.event.delta.content, ToolCall
|
||||
): # first chunk may contain entire call
|
||||
assert first.event.delta.parse_status == ToolCallParseStatus.started
|
||||
|
||||
last = grouped[ChatCompletionResponseEventType.progress][-1]
|
||||
# assert last.event.stop_reason == expected_stop_reason
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue