forked from phoenix-oss/llama-stack-mirror
# What does this PR do? - To make it easier, delete existing `eval/scoring/scoring_function` apis. There will be a bunch of broken impls here. The sequence is: 1. migrate benchmark graders 2. clean up existing scoring functions - Add a skeleton evaluation impl to make tests pass. ## Test Plan tested in following PRs [//]: # (## Documentation) |
||
---|---|---|
.. | ||
__init__.py | ||
agents.py | ||
datasetio.py | ||
evaluation.py | ||
inference.py | ||
post_training.py | ||
safety.py | ||
telemetry.py | ||
tool_runtime.py | ||
vector_io.py |