forked from phoenix-oss/llama-stack-mirror
openapi gen
This commit is contained in:
parent
24d48b3692
commit
205a50f10b
2 changed files with 9 additions and 5 deletions
2
docs/_static/llama-stack-spec.html
vendored
2
docs/_static/llama-stack-spec.html
vendored
|
@ -8548,7 +8548,7 @@
|
||||||
},
|
},
|
||||||
"additionalProperties": false,
|
"additionalProperties": false,
|
||||||
"title": "EvaluationTask",
|
"title": "EvaluationTask",
|
||||||
"description": "A task for evaluation. To specify a task, one of the following must be provided: - `benchmark_id`: Run evaluation task against a benchmark_id - `dataset_id` and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids - `data_source` and `grader_ids`: Run evaluation task against a data source (e.g. rows, uri, etc.) and a list of grader_ids"
|
"description": "A task for evaluation. To specify a task, one of the following must be provided: - `benchmark_id`: Run evaluation task against a benchmark_id. Use this when you have a curated dataset and have settled on the graders. - `dataset_id` and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids. Use this when you have datasets and / or are iterating on your graders. - `data_source` and `grader_ids`: Run evaluation task against a data source (e.g. rows, uri, etc.) and a list of grader_ids. Prefer this when you are early in your evaluation cycle and experimenting much more with your data and graders."
|
||||||
},
|
},
|
||||||
"GradeRequest": {
|
"GradeRequest": {
|
||||||
"type": "object",
|
"type": "object",
|
||||||
|
|
12
docs/_static/llama-stack-spec.yaml
vendored
12
docs/_static/llama-stack-spec.yaml
vendored
|
@ -5924,10 +5924,14 @@ components:
|
||||||
title: EvaluationTask
|
title: EvaluationTask
|
||||||
description: >-
|
description: >-
|
||||||
A task for evaluation. To specify a task, one of the following must be provided:
|
A task for evaluation. To specify a task, one of the following must be provided:
|
||||||
- `benchmark_id`: Run evaluation task against a benchmark_id - `dataset_id`
|
- `benchmark_id`: Run evaluation task against a benchmark_id. Use this when
|
||||||
and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids
|
you have a curated dataset and have settled on the graders. - `dataset_id`
|
||||||
- `data_source` and `grader_ids`: Run evaluation task against a data source
|
and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids.
|
||||||
(e.g. rows, uri, etc.) and a list of grader_ids
|
Use this when you have datasets and / or are iterating on your graders. -
|
||||||
|
`data_source` and `grader_ids`: Run evaluation task against a data source
|
||||||
|
(e.g. rows, uri, etc.) and a list of grader_ids. Prefer this when you are
|
||||||
|
early in your evaluation cycle and experimenting much more with your data
|
||||||
|
and graders.
|
||||||
GradeRequest:
|
GradeRequest:
|
||||||
type: object
|
type: object
|
||||||
properties:
|
properties:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue