mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 20:12:33 +00:00
simplified some, walked back some decisions
This commit is contained in:
parent
af7472cdb0
commit
636764c2a1
90 changed files with 887 additions and 570 deletions
|
|
@ -41,11 +41,10 @@ class PromptServiceImpl(Prompts):
|
|||
|
||||
async def initialize(self) -> None:
|
||||
# Use metadata store backend with prompts-specific namespace
|
||||
metadata_ref = self.config.run_config.storage.metadata
|
||||
prompts_ref = KVStoreReference(
|
||||
namespace="prompts",
|
||||
backend=metadata_ref.backend if metadata_ref else None,
|
||||
)
|
||||
metadata_ref = self.config.run_config.metadata_store
|
||||
if not metadata_ref:
|
||||
raise ValueError("metadata_store must be configured in run config")
|
||||
prompts_ref = KVStoreReference(namespace="prompts", backend=metadata_ref.backend)
|
||||
self.kvstore = await kvstore_impl(prompts_ref)
|
||||
|
||||
def _get_default_key(self, prompt_id: str) -> str:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue