diff --git a/tests/integration/agents/test_openai_responses.py b/tests/integration/agents/test_openai_responses.py index c9c1d4fa8..a3a255144 100644 --- a/tests/integration/agents/test_openai_responses.py +++ b/tests/integration/agents/test_openai_responses.py @@ -41,6 +41,7 @@ def openai_client(client_with_models): ], ], ) +@pytest.mark.skip(reason="Very flaky, sometimes there is a message not a function call, standard tool calling issues") def test_responses_store(openai_client, client_with_models, text_model_id, stream, tools): if isinstance(client_with_models, LlamaStackAsLibraryClient): pytest.skip("OpenAI responses are not supported when testing with library client yet.") diff --git a/tests/integration/inference/test_openai_completion.py b/tests/integration/inference/test_openai_completion.py index 6f8a05a45..28121ad44 100644 --- a/tests/integration/inference/test_openai_completion.py +++ b/tests/integration/inference/test_openai_completion.py @@ -274,6 +274,7 @@ def test_inference_store(openai_client, client_with_models, text_model_id, strea False, ], ) +@pytest.mark.skip(reason="Very flaky, tool calling really wacky on CI") def test_inference_store_tool_calls(openai_client, client_with_models, text_model_id, stream): skip_if_model_doesnt_support_openai_chat_completion(client_with_models, text_model_id) client = openai_client