From d3956a1d22bbd480f4e14fd3f79b01cab7a23661 Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Mon, 25 Nov 2024 22:02:45 -0800 Subject: [PATCH] fix description --- .../scoring/braintrust/scoring_fn/fn_defs/answer_correctness.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_stack/providers/inline/scoring/braintrust/scoring_fn/fn_defs/answer_correctness.py b/llama_stack/providers/inline/scoring/braintrust/scoring_fn/fn_defs/answer_correctness.py index 554590f12..dc5df8e78 100644 --- a/llama_stack/providers/inline/scoring/braintrust/scoring_fn/fn_defs/answer_correctness.py +++ b/llama_stack/providers/inline/scoring/braintrust/scoring_fn/fn_defs/answer_correctness.py @@ -10,7 +10,7 @@ from llama_stack.apis.scoring_functions import ScoringFn answer_correctness_fn_def = ScoringFn( identifier="braintrust::answer-correctness", - description="Test whether an output is factual, compared to an original (`expected`) value. One of Braintrust LLM basd scorer https://github.com/braintrustdata/autoevals/blob/main/py/autoevals/llm.py", + description="Scores the correctness of the answer based on the ground truth.. One of Braintrust LLM basd scorer https://github.com/braintrustdata/autoevals/blob/main/py/autoevals/llm.py", params=None, provider_id="braintrust", provider_resource_id="answer-correctness",