mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
test: Enable test_text_chat_completion_with_tool_choice_required for remote::vllm (#1148)
This commit is contained in:
parent
8de7cf103b
commit
a66b4c4c81
1 changed files with 0 additions and 2 deletions
|
@ -250,8 +250,6 @@ def test_text_chat_completion_with_tool_calling_and_streaming(
|
||||||
def test_text_chat_completion_with_tool_choice_required(
|
def test_text_chat_completion_with_tool_choice_required(
|
||||||
llama_stack_client, text_model_id, get_weather_tool_definition, provider_tool_format, inference_provider_type
|
llama_stack_client, text_model_id, get_weather_tool_definition, provider_tool_format, inference_provider_type
|
||||||
):
|
):
|
||||||
if inference_provider_type == "remote::vllm":
|
|
||||||
pytest.xfail("vllm-project/vllm#13002")
|
|
||||||
response = llama_stack_client.inference.chat_completion(
|
response = llama_stack_client.inference.chat_completion(
|
||||||
model_id=text_model_id,
|
model_id=text_model_id,
|
||||||
messages=[
|
messages=[
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue