mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-16 08:49:28 +00:00
minimize change
This commit is contained in:
parent
ed78090b8e
commit
06d02bf3de
23 changed files with 161 additions and 197 deletions
|
|
@ -314,8 +314,7 @@ class MetaReferenceAgentsImpl(Agents):
|
|||
return paginate_records(session_dicts, start_index, limit)
|
||||
|
||||
async def shutdown(self) -> None:
|
||||
await self.persistence_store.close()
|
||||
await self.responses_store.shutdown()
|
||||
pass
|
||||
|
||||
# OpenAI responses
|
||||
async def get_openai_response(
|
||||
|
|
|
|||
|
|
@ -129,7 +129,6 @@ class ReferenceBatchesImpl(Batches):
|
|||
# don't cancel tasks - just let them stop naturally on shutdown
|
||||
# cancelling would mark batches as "cancelled" in the database
|
||||
logger.info(f"Shutdown initiated with {len(self._processing_tasks)} active batch processing tasks")
|
||||
await self.kvstore.close()
|
||||
|
||||
# TODO (SECURITY): this currently works w/ configured api keys, not with x-llamastack-provider-data or with user policy restrictions
|
||||
async def create_batch(
|
||||
|
|
|
|||
|
|
@ -64,8 +64,7 @@ class MetaReferenceEvalImpl(
|
|||
benchmark = Benchmark.model_validate_json(benchmark)
|
||||
self.benchmarks[benchmark.identifier] = benchmark
|
||||
|
||||
async def shutdown(self) -> None:
|
||||
await self.kvstore.close()
|
||||
async def shutdown(self) -> None: ...
|
||||
|
||||
async def register_benchmark(self, task_def: Benchmark) -> None:
|
||||
# Store in kvstore
|
||||
|
|
|
|||
|
|
@ -62,8 +62,7 @@ class LocalfsFilesImpl(Files):
|
|||
)
|
||||
|
||||
async def shutdown(self) -> None:
|
||||
if self.sql_store:
|
||||
await self.sql_store.close()
|
||||
pass
|
||||
|
||||
def _generate_file_id(self) -> str:
|
||||
"""Generate a unique file ID for OpenAI API."""
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue