llama-stack/llama_stack/providers/inline
Botao Chen b751f7003d
feat: add aggregation_functions to llm_as_judge_405b_simpleqa (#1164)
as title, to let scoring function llm_as_judge_405b_simpleqa output
aggregated_results.

We can leverage categorical_count to calculate the % of correctness as
eval benchmark metrics
2025-02-19 19:42:04 -08:00
..
agents feat: log start, complete time to Agent steps (#1116) 2025-02-14 17:48:06 -08:00
datasetio build: format codebase imports using ruff linter (#1028) 2025-02-13 10:06:21 -08:00
eval build: configure ruff from pyproject.toml (#1100) 2025-02-14 09:01:57 -08:00
inference chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
ios/inference LocalInferenceImpl update for LS 0.1 (#911) 2025-02-02 09:49:40 -08:00
post_training fix: miscellaneous job management improvements in torchtune (#1136) 2025-02-19 19:09:37 -08:00
safety chore: move all Llama Stack types from llama-models to llama-stack (#1098) 2025-02-14 09:10:59 -08:00
scoring feat: add aggregation_functions to llm_as_judge_405b_simpleqa (#1164) 2025-02-19 19:42:04 -08:00
telemetry build: format codebase imports using ruff linter (#1028) 2025-02-13 10:06:21 -08:00
tool_runtime fix: Get distro_codegen.py working with default deps and enabled in pre-commit hooks (#1123) 2025-02-19 18:39:20 -08:00
vector_io feat: Chunk sqlite-vec writes (#1094) 2025-02-19 19:07:46 -08:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00