Resolved merge conflicts

This commit is contained in:
Chantal D Gama Rose 2025-03-14 14:18:23 -07:00
parent 3b3195d8e6
commit eebb0f6e61
8 changed files with 0 additions and 124 deletions

View file

@ -102,4 +102,3 @@ def pytest_generate_tests(metafunc):
get_provider_fixture_overrides(metafunc.config, available_fixtures) or DEFAULT_PROVIDER_COMBINATIONS
)
metafunc.parametrize("safety_stack", combinations, indirect=True)

View file

@ -26,18 +26,11 @@ providers:
namespace: null
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/nvidia}/faiss_store.db
safety:
<<<<<<< HEAD
- provider_id: nvidia
provider_type: remote::nvidia
config:
guardrails_service_url: ${env.GUARDRAILS_SERVICE_URL:http://localhost:7331}
config_id: self-check
=======
- provider_id: llama-guard
provider_type: inline::llama-guard
config:
excluded_categories: []
>>>>>>> upstream/main
agents:
- provider_id: meta-reference
provider_type: inline::meta-reference
@ -62,16 +55,6 @@ providers:
namespace: null
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/nvidia}/meta_reference_eval.db
datasetio:
<<<<<<< HEAD
=======
- provider_id: huggingface
provider_type: remote::huggingface
config:
kvstore:
type: sqlite
namespace: null
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/nvidia}/huggingface_datasetio.db
>>>>>>> upstream/main
- provider_id: localfs
provider_type: inline::localfs
config:

View file

@ -3,8 +3,5 @@
#
# This source code is licensed under the terms described in the LICENSE file in
# the root directory of this source tree.
<<<<<<<< HEAD:llama_stack/templates/open-benchmark/__init__.py
from .open_benchmark import get_distribution_template # noqa: F401
========
>>>>>>>> upstream/main:llama_stack/providers/inline/scoring/basic/utils/bfcl/__init__.py

View file

@ -226,8 +226,6 @@ def get_distribution_template() -> DistributionTemplate:
"chat_completion_input": {"type": "string"},
},
),
<<<<<<< HEAD
=======
DatasetInput(
dataset_id="bfcl",
provider_id="huggingface",
@ -244,7 +242,6 @@ def get_distribution_template() -> DistributionTemplate:
"chat_completion_input": {"type": "string"},
},
),
>>>>>>> upstream/main
]
default_benchmarks = [
@ -268,14 +265,11 @@ def get_distribution_template() -> DistributionTemplate:
dataset_id="math_500",
scoring_functions=["basic::regex_parser_math_response"],
),
<<<<<<< HEAD
=======
BenchmarkInput(
benchmark_id="meta-reference-bfcl",
dataset_id="bfcl",
scoring_functions=["basic::bfcl"],
),
>>>>>>> upstream/main
]
return DistributionTemplate(
name=name,

View file

@ -216,8 +216,6 @@ datasets:
split: test
dataset_id: math_500
provider_id: huggingface
<<<<<<< HEAD
=======
- dataset_schema:
function:
type: string
@ -236,7 +234,6 @@ datasets:
split: train
dataset_id: bfcl
provider_id: huggingface
>>>>>>> upstream/main
scoring_fns: []
benchmarks:
- dataset_id: simpleqa
@ -259,14 +256,11 @@ benchmarks:
- basic::regex_parser_math_response
metadata: {}
benchmark_id: meta-reference-math-500
<<<<<<< HEAD
=======
- dataset_id: bfcl
scoring_functions:
- basic::bfcl
metadata: {}
benchmark_id: meta-reference-bfcl
>>>>>>> upstream/main
tool_groups:
- toolgroup_id: builtin::websearch
provider_id: tavily-search