# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the terms described in the LICENSE file in # the root directory of this source tree. import pytest from ..conftest import get_provider_fixture_overrides from ..inference.fixtures import INFERENCE_FIXTURES from .fixtures import MEMORY_FIXTURES DEFAULT_PROVIDER_COMBINATIONS = [ pytest.param( { "inference": "meta_reference", "memory": "faiss", }, id="meta_reference", marks=pytest.mark.meta_reference, ), pytest.param( { "inference": "ollama", "memory": "pgvector", }, id="ollama", marks=pytest.mark.ollama, ), pytest.param( { "inference": "together", "memory": "chroma", }, id="chroma", marks=pytest.mark.chroma, ), pytest.param( { "inference": "bedrock", "memory": "qdrant", }, id="qdrant", marks=pytest.mark.qdrant, ), pytest.param( { "inference": "fireworks", "memory": "weaviate", }, id="weaviate", marks=pytest.mark.weaviate, ), ] def pytest_addoption(parser): parser.addoption( "--inference-model", action="store", default=None, help="Specify the inference model to use for testing", ) def pytest_configure(config): for fixture_name in MEMORY_FIXTURES: config.addinivalue_line( "markers", f"{fixture_name}: marks tests as {fixture_name} specific", ) def pytest_generate_tests(metafunc): if "inference_model" in metafunc.fixturenames: model = metafunc.config.getoption("--inference-model") if not model: raise ValueError( "No inference model specified. Please provide a valid inference model." ) params = [pytest.param(model, id="")] metafunc.parametrize("inference_model", params, indirect=True) if "memory_stack" in metafunc.fixturenames: available_fixtures = { "inference": INFERENCE_FIXTURES, "memory": MEMORY_FIXTURES, } combinations = ( get_provider_fixture_overrides(metafunc.config, available_fixtures) or DEFAULT_PROVIDER_COMBINATIONS ) metafunc.parametrize("memory_stack", combinations, indirect=True)