mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-23 03:42:26 +00:00
refactor component path
Signed-off-by: Francisco Javier Arceo <farceo@redhat.com>
This commit is contained in:
parent
a29b62f820
commit
19d2555424
116 changed files with 7064 additions and 1265 deletions
|
|
@ -22,6 +22,15 @@ logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(
|
|||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
skip_because_resource_intensive = pytest.mark.skip(
|
||||
reason="""
|
||||
Post training tests are extremely resource intensive. They download large models and partly as a result,
|
||||
are very slow to run. We cannot run them on every single PR update. CI should be considered
|
||||
a scarce resource and properly utilitized.
|
||||
"""
|
||||
)
|
||||
|
||||
|
||||
@pytest.fixture(autouse=True)
|
||||
def capture_output(capsys):
|
||||
"""Fixture to capture and display output during test execution."""
|
||||
|
|
@ -57,6 +66,7 @@ class TestPostTraining:
|
|||
],
|
||||
)
|
||||
@pytest.mark.timeout(360) # 6 minutes timeout
|
||||
@skip_because_resource_intensive
|
||||
def test_supervised_fine_tune(self, llama_stack_client, purpose, source):
|
||||
logger.info("Starting supervised fine-tuning test")
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue