forked from phoenix-oss/llama-stack-mirror
openapi gen
This commit is contained in:
parent
24d48b3692
commit
205a50f10b
2 changed files with 9 additions and 5 deletions
2
docs/_static/llama-stack-spec.html
vendored
2
docs/_static/llama-stack-spec.html
vendored
|
@ -8548,7 +8548,7 @@
|
|||
},
|
||||
"additionalProperties": false,
|
||||
"title": "EvaluationTask",
|
||||
"description": "A task for evaluation. To specify a task, one of the following must be provided: - `benchmark_id`: Run evaluation task against a benchmark_id - `dataset_id` and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids - `data_source` and `grader_ids`: Run evaluation task against a data source (e.g. rows, uri, etc.) and a list of grader_ids"
|
||||
"description": "A task for evaluation. To specify a task, one of the following must be provided: - `benchmark_id`: Run evaluation task against a benchmark_id. Use this when you have a curated dataset and have settled on the graders. - `dataset_id` and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids. Use this when you have datasets and / or are iterating on your graders. - `data_source` and `grader_ids`: Run evaluation task against a data source (e.g. rows, uri, etc.) and a list of grader_ids. Prefer this when you are early in your evaluation cycle and experimenting much more with your data and graders."
|
||||
},
|
||||
"GradeRequest": {
|
||||
"type": "object",
|
||||
|
|
12
docs/_static/llama-stack-spec.yaml
vendored
12
docs/_static/llama-stack-spec.yaml
vendored
|
@ -5924,10 +5924,14 @@ components:
|
|||
title: EvaluationTask
|
||||
description: >-
|
||||
A task for evaluation. To specify a task, one of the following must be provided:
|
||||
- `benchmark_id`: Run evaluation task against a benchmark_id - `dataset_id`
|
||||
and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids
|
||||
- `data_source` and `grader_ids`: Run evaluation task against a data source
|
||||
(e.g. rows, uri, etc.) and a list of grader_ids
|
||||
- `benchmark_id`: Run evaluation task against a benchmark_id. Use this when
|
||||
you have a curated dataset and have settled on the graders. - `dataset_id`
|
||||
and `grader_ids`: Run evaluation task against a dataset_id and a list of grader_ids.
|
||||
Use this when you have datasets and / or are iterating on your graders. -
|
||||
`data_source` and `grader_ids`: Run evaluation task against a data source
|
||||
(e.g. rows, uri, etc.) and a list of grader_ids. Prefer this when you are
|
||||
early in your evaluation cycle and experimenting much more with your data
|
||||
and graders.
|
||||
GradeRequest:
|
||||
type: object
|
||||
properties:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue