forked from phoenix-oss/llama-stack-mirror
Update llama_stack/apis/evaluation/evaluation.py
Co-authored-by: Ashwin Bharambe <ashwin.bharambe@gmail.com>
This commit is contained in:
parent
d994499f09
commit
85cad639ca
1 changed files with 1 additions and 1 deletions
|
@ -51,7 +51,7 @@ EvaluationCandidate = register_schema(
|
|||
class EvaluationTask(BaseModel):
|
||||
"""
|
||||
A task for evaluation. To specify a task, one of the following must be provided:
|
||||
- `benchmark_id`: Run evaluation task against a benchmark_id
|
||||
- `benchmark_id`: Run evaluation task against a benchmark_id. Use this when you have a curated dataset and have settled on the graders.
|
||||
- `dataset_id` and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids
|
||||
- `data_source` and `grader_ids`: Run evaluation task against a data source (e.g. rows, uri, etc.) and a list of grader_ids
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue