fix: disable test_responses_store (#2244)

The test depends on llama's tool calling ability. In the CI, we run with
a small ollama model.

The fix might be to check for either message or function_call because
the model is flaky and we aren't really testing that behavior?
This commit is contained in:
Ashwin Bharambe 2025-05-24 08:18:06 -07:00 committed by GitHub
parent 84751f3e55
commit 66f09f24ed
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2 changed files with 2 additions and 0 deletions

View file

@ -41,6 +41,7 @@ def openai_client(client_with_models):
],
],
)
@pytest.mark.skip(reason="Very flaky, sometimes there is a message not a function call, standard tool calling issues")
def test_responses_store(openai_client, client_with_models, text_model_id, stream, tools):
if isinstance(client_with_models, LlamaStackAsLibraryClient):
pytest.skip("OpenAI responses are not supported when testing with library client yet.")

View file

@ -274,6 +274,7 @@ def test_inference_store(openai_client, client_with_models, text_model_id, strea
False,
],
)
@pytest.mark.skip(reason="Very flaky, tool calling really wacky on CI")
def test_inference_store_tool_calls(openai_client, client_with_models, text_model_id, stream):
skip_if_model_doesnt_support_openai_chat_completion(client_with_models, text_model_id)
client = openai_client