mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-10 04:08:31 +00:00
test: Enable test_text_chat_completion_with_tool_choice_required for remote::vllm
Signed-off-by: Yuan Tang <terrytangyuan@gmail.com>
This commit is contained in:
parent
8de7cf103b
commit
eb3541c9fb
1 changed files with 0 additions and 2 deletions
|
@ -250,8 +250,6 @@ def test_text_chat_completion_with_tool_calling_and_streaming(
|
|||
def test_text_chat_completion_with_tool_choice_required(
|
||||
llama_stack_client, text_model_id, get_weather_tool_definition, provider_tool_format, inference_provider_type
|
||||
):
|
||||
if inference_provider_type == "remote::vllm":
|
||||
pytest.xfail("vllm-project/vllm#13002")
|
||||
response = llama_stack_client.inference.chat_completion(
|
||||
model_id=text_model_id,
|
||||
messages=[
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue