mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-08 19:10:56 +00:00
wip
This commit is contained in:
parent
4764762dd4
commit
5b7d24b1c3
3 changed files with 14 additions and 16 deletions
|
|
@ -14,10 +14,6 @@ from llama_stack.providers.impls.meta_reference.evals.datas.dataset_registry imp
|
|||
get_dataset,
|
||||
)
|
||||
|
||||
# from llama_stack.providers.impls.meta_reference.evals.tasks.task_registry import (
|
||||
# get_task,
|
||||
# )
|
||||
|
||||
from .config import MetaReferenceEvalsImplConfig
|
||||
|
||||
|
||||
|
|
@ -45,7 +41,9 @@ class MetaReferenceEvalsImpl(Evals):
|
|||
|
||||
# TODO: replace w/ batch inference & async return eval job
|
||||
generation_outputs = []
|
||||
print("generation start")
|
||||
for msg in x1[:5]:
|
||||
print("generation for msg: ", msg)
|
||||
response = self.inference_api.chat_completion(
|
||||
model=model,
|
||||
messages=[msg],
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue