diff --git a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb index 2861c8499..8eecf84ab 100644 --- a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb +++ b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb @@ -1214,7 +1214,7 @@ " \"sampling_params\": {\n", " \"strategy\": {\n", " \"type\": \"greedy\",\n", - " },b\n", + " },\n", " \"max_tokens\": 4096,\n", " \"repeat_penalty\": 1.0,\n", " },\n", diff --git a/llama_stack/distribution/routers/routing_tables.py b/llama_stack/distribution/routers/routing_tables.py index 563c5c5ab..ec258af49 100644 --- a/llama_stack/distribution/routers/routing_tables.py +++ b/llama_stack/distribution/routers/routing_tables.py @@ -472,20 +472,16 @@ class BenchmarksRoutingTable(CommonRoutingTableImpl, Benchmarks): async def DEPRECATED_list_eval_tasks(self) -> ListBenchmarksResponse: logger.warning("DEPRECATED: Use /eval/benchmarks instead") return await self.list_benchmarks() - return await self.list_benchmarks() async def DEPRECATED_get_eval_task( self, - task_id: str, eval_task_id: str, ) -> Optional[Benchmark]: logger.warning("DEPRECATED: Use /eval/benchmarks instead") - return await self.get_benchmark(task_id) return await self.get_benchmark(eval_task_id) async def DEPRECATED_register_eval_task( self, - task_id: str, eval_task_id: str, dataset_id: str, scoring_functions: List[str], @@ -494,8 +490,6 @@ class BenchmarksRoutingTable(CommonRoutingTableImpl, Benchmarks): metadata: Optional[Dict[str, Any]] = None, ) -> None: logger.warning("DEPRECATED: Use /eval/benchmarks instead") - return await self.register_benchmark( - benchmark_id=task_id, return await self.register_benchmark( benchmark_id=eval_task_id, dataset_id=dataset_id,