diff --git a/llama_stack/templates/open-benchmark/open_benchmark.py b/llama_stack/templates/open-benchmark/open_benchmark.py index b019b2e4f..d1c27e901 100644 --- a/llama_stack/templates/open-benchmark/open_benchmark.py +++ b/llama_stack/templates/open-benchmark/open_benchmark.py @@ -205,7 +205,6 @@ def get_distribution_template() -> DistributionTemplate: ), DatasetInput( dataset_id="docvqa", - provider_id="huggingface", purpose=DatasetPurpose.eval_messages_answer, source=URIDataSource( uri="huggingface://datasets/llamastack/docvqa?split=val", diff --git a/llama_stack/templates/open-benchmark/run.yaml b/llama_stack/templates/open-benchmark/run.yaml index 4233bc4f5..80a517fe8 100644 --- a/llama_stack/templates/open-benchmark/run.yaml +++ b/llama_stack/templates/open-benchmark/run.yaml @@ -188,17 +188,12 @@ datasets: uri: huggingface://datasets/llamastack/bfcl_v3?split=train metadata: {} dataset_id: bfcl -<<<<<<< HEAD -======= - provider_id: huggingface - purpose: eval/messages-answer source: type: uri uri: huggingface://datasets/llamastack/docvqa?split=val metadata: {} dataset_id: docvqa - provider_id: huggingface ->>>>>>> 81df5a4e (add DocVQA scoring) scoring_fns: [] benchmarks: - dataset_id: simpleqa