mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-01 16:24:44 +00:00
registery to handle updates and deletes
This commit is contained in:
parent
4b1b196251
commit
9e68ed3f36
5 changed files with 113 additions and 120 deletions
|
@ -21,7 +21,7 @@
|
||||||
"info": {
|
"info": {
|
||||||
"title": "[DRAFT] Llama Stack Specification",
|
"title": "[DRAFT] Llama Stack Specification",
|
||||||
"version": "0.0.1",
|
"version": "0.0.1",
|
||||||
"description": "This is the specification of the llama stack that provides\n a set of endpoints and their corresponding interfaces that are tailored to\n best leverage Llama Models. The specification is still in draft and subject to change.\n Generated at 2024-11-13 15:29:27.077633"
|
"description": "This is the specification of the llama stack that provides\n a set of endpoints and their corresponding interfaces that are tailored to\n best leverage Llama Models. The specification is still in draft and subject to change.\n Generated at 2024-11-13 18:16:59.065989"
|
||||||
},
|
},
|
||||||
"servers": [
|
"servers": [
|
||||||
{
|
{
|
||||||
|
@ -7965,10 +7965,7 @@
|
||||||
],
|
],
|
||||||
"tags": [
|
"tags": [
|
||||||
{
|
{
|
||||||
"name": "Datasets"
|
"name": "Shields"
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "Inference"
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "ScoringFunctions"
|
"name": "ScoringFunctions"
|
||||||
|
@ -7977,20 +7974,38 @@
|
||||||
"name": "MemoryBanks"
|
"name": "MemoryBanks"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Telemetry"
|
"name": "Datasets"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "PostTraining"
|
"name": "Agents"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Models"
|
"name": "DatasetIO"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "Inference"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Inspect"
|
"name": "Inspect"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"name": "Memory"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "Models"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "PostTraining"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"name": "Safety"
|
"name": "Safety"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"name": "SyntheticDataGeneration"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "EvalTasks"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"name": "Scoring"
|
"name": "Scoring"
|
||||||
},
|
},
|
||||||
|
@ -8001,22 +8016,7 @@
|
||||||
"name": "Eval"
|
"name": "Eval"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "SyntheticDataGeneration"
|
"name": "Telemetry"
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "EvalTasks"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "Shields"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "Memory"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "DatasetIO"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "Agents"
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "BuiltinTool",
|
"name": "BuiltinTool",
|
||||||
|
|
|
@ -3414,7 +3414,7 @@ info:
|
||||||
description: "This is the specification of the llama stack that provides\n \
|
description: "This is the specification of the llama stack that provides\n \
|
||||||
\ a set of endpoints and their corresponding interfaces that are tailored\
|
\ a set of endpoints and their corresponding interfaces that are tailored\
|
||||||
\ to\n best leverage Llama Models. The specification is still in\
|
\ to\n best leverage Llama Models. The specification is still in\
|
||||||
\ draft and subject to change.\n Generated at 2024-11-13 15:29:27.077633"
|
\ draft and subject to change.\n Generated at 2024-11-13 18:16:59.065989"
|
||||||
title: '[DRAFT] Llama Stack Specification'
|
title: '[DRAFT] Llama Stack Specification'
|
||||||
version: 0.0.1
|
version: 0.0.1
|
||||||
jsonSchemaDialect: https://json-schema.org/draft/2020-12/schema
|
jsonSchemaDialect: https://json-schema.org/draft/2020-12/schema
|
||||||
|
@ -4824,24 +4824,24 @@ security:
|
||||||
servers:
|
servers:
|
||||||
- url: http://any-hosted-llama-stack.com
|
- url: http://any-hosted-llama-stack.com
|
||||||
tags:
|
tags:
|
||||||
- name: Datasets
|
- name: Shields
|
||||||
- name: Inference
|
|
||||||
- name: ScoringFunctions
|
- name: ScoringFunctions
|
||||||
- name: MemoryBanks
|
- name: MemoryBanks
|
||||||
- name: Telemetry
|
- name: Datasets
|
||||||
- name: PostTraining
|
- name: Agents
|
||||||
- name: Models
|
- name: DatasetIO
|
||||||
|
- name: Inference
|
||||||
- name: Inspect
|
- name: Inspect
|
||||||
|
- name: Memory
|
||||||
|
- name: Models
|
||||||
|
- name: PostTraining
|
||||||
- name: Safety
|
- name: Safety
|
||||||
|
- name: SyntheticDataGeneration
|
||||||
|
- name: EvalTasks
|
||||||
- name: Scoring
|
- name: Scoring
|
||||||
- name: BatchInference
|
- name: BatchInference
|
||||||
- name: Eval
|
- name: Eval
|
||||||
- name: SyntheticDataGeneration
|
- name: Telemetry
|
||||||
- name: EvalTasks
|
|
||||||
- name: Shields
|
|
||||||
- name: Memory
|
|
||||||
- name: DatasetIO
|
|
||||||
- name: Agents
|
|
||||||
- description: <SchemaDefinition schemaRef="#/components/schemas/BuiltinTool" />
|
- description: <SchemaDefinition schemaRef="#/components/schemas/BuiltinTool" />
|
||||||
name: BuiltinTool
|
name: BuiltinTool
|
||||||
- description: <SchemaDefinition schemaRef="#/components/schemas/CompletionMessage"
|
- description: <SchemaDefinition schemaRef="#/components/schemas/CompletionMessage"
|
||||||
|
|
|
@ -64,5 +64,5 @@ class Models(Protocol):
|
||||||
metadata: Optional[Dict[str, Any]] = None,
|
metadata: Optional[Dict[str, Any]] = None,
|
||||||
) -> Model: ...
|
) -> Model: ...
|
||||||
|
|
||||||
@webmethod(route="/models/delete", method="DELETE")
|
@webmethod(route="/models/delete", method="POST")
|
||||||
async def delete_model(self, model_id: str) -> None: ...
|
async def delete_model(self, model_id: str) -> None: ...
|
||||||
|
|
|
@ -124,8 +124,8 @@ class CommonRoutingTableImpl(RoutingTable):
|
||||||
apiname, objtype = apiname_object()
|
apiname, objtype = apiname_object()
|
||||||
|
|
||||||
# Get objects from disk registry
|
# Get objects from disk registry
|
||||||
objects = self.dist_registry.get_cached(objtype, routing_key)
|
obj = self.dist_registry.get_cached(objtype, routing_key)
|
||||||
if not objects:
|
if not obj:
|
||||||
provider_ids = list(self.impls_by_provider_id.keys())
|
provider_ids = list(self.impls_by_provider_id.keys())
|
||||||
if len(provider_ids) > 1:
|
if len(provider_ids) > 1:
|
||||||
provider_ids_str = f"any of the providers: {', '.join(provider_ids)}"
|
provider_ids_str = f"any of the providers: {', '.join(provider_ids)}"
|
||||||
|
@ -135,9 +135,8 @@ class CommonRoutingTableImpl(RoutingTable):
|
||||||
f"{objtype.capitalize()} `{routing_key}` not served by {provider_ids_str}. Make sure there is an {apiname} provider serving this {objtype}."
|
f"{objtype.capitalize()} `{routing_key}` not served by {provider_ids_str}. Make sure there is an {apiname} provider serving this {objtype}."
|
||||||
)
|
)
|
||||||
|
|
||||||
for obj in objects:
|
if not provider_id or provider_id == obj.provider_id:
|
||||||
if not provider_id or provider_id == obj.provider_id:
|
return self.impls_by_provider_id[obj.provider_id]
|
||||||
return self.impls_by_provider_id[obj.provider_id]
|
|
||||||
|
|
||||||
raise ValueError(f"Provider not found for `{routing_key}`")
|
raise ValueError(f"Provider not found for `{routing_key}`")
|
||||||
|
|
||||||
|
@ -145,30 +144,36 @@ class CommonRoutingTableImpl(RoutingTable):
|
||||||
self, type: str, identifier: str
|
self, type: str, identifier: str
|
||||||
) -> Optional[RoutableObjectWithProvider]:
|
) -> Optional[RoutableObjectWithProvider]:
|
||||||
# Get from disk registry
|
# Get from disk registry
|
||||||
objects = await self.dist_registry.get(type, identifier)
|
obj = await self.dist_registry.get(type, identifier)
|
||||||
if not objects:
|
if not obj:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
assert len(objects) == 1
|
return obj
|
||||||
return objects[0]
|
|
||||||
|
|
||||||
async def delete_object(self, obj: RoutableObjectWithProvider) -> None:
|
async def delete_object(self, obj: RoutableObjectWithProvider) -> None:
|
||||||
await self.dist_registry.delete(obj.type, obj.identifier)
|
await self.dist_registry.delete(obj.type, obj.identifier)
|
||||||
# TODO: delete from provider
|
# TODO: delete from provider
|
||||||
|
|
||||||
|
async def update_object(
|
||||||
|
self, obj: RoutableObjectWithProvider
|
||||||
|
) -> RoutableObjectWithProvider:
|
||||||
|
registered_obj = await register_object_with_provider(
|
||||||
|
obj, self.impls_by_provider_id[obj.provider_id]
|
||||||
|
)
|
||||||
|
return await self.dist_registry.update(registered_obj)
|
||||||
|
|
||||||
async def register_object(
|
async def register_object(
|
||||||
self, obj: RoutableObjectWithProvider
|
self, obj: RoutableObjectWithProvider
|
||||||
) -> RoutableObjectWithProvider:
|
) -> RoutableObjectWithProvider:
|
||||||
# Get existing objects from registry
|
# Get existing objects from registry
|
||||||
existing_objects = await self.dist_registry.get(obj.type, obj.identifier)
|
existing_obj = await self.dist_registry.get(obj.type, obj.identifier)
|
||||||
|
|
||||||
# Check for existing registration
|
# Check for existing registration
|
||||||
for existing_obj in existing_objects:
|
if existing_obj and existing_obj.provider_id == obj.provider_id:
|
||||||
if existing_obj.provider_id == obj.provider_id or not obj.provider_id:
|
print(
|
||||||
print(
|
f"`{obj.identifier}` already registered with `{existing_obj.provider_id}`"
|
||||||
f"`{obj.identifier}` already registered with `{existing_obj.provider_id}`"
|
)
|
||||||
)
|
return existing_obj
|
||||||
return existing_obj
|
|
||||||
|
|
||||||
# if provider_id is not specified, pick an arbitrary one from existing entries
|
# if provider_id is not specified, pick an arbitrary one from existing entries
|
||||||
if not obj.provider_id and len(self.impls_by_provider_id) > 0:
|
if not obj.provider_id and len(self.impls_by_provider_id) > 0:
|
||||||
|
@ -247,7 +252,7 @@ class ModelsRoutingTable(CommonRoutingTableImpl, Models):
|
||||||
provider_id=provider_id or existing_model.provider_id,
|
provider_id=provider_id or existing_model.provider_id,
|
||||||
metadata=metadata or existing_model.metadata,
|
metadata=metadata or existing_model.metadata,
|
||||||
)
|
)
|
||||||
registered_model = await self.register_object(updated_model)
|
registered_model = await self.update_object(updated_model)
|
||||||
return registered_model
|
return registered_model
|
||||||
|
|
||||||
async def delete_model(self, model_id: str) -> None:
|
async def delete_model(self, model_id: str) -> None:
|
||||||
|
|
|
@ -26,9 +26,13 @@ class DistributionRegistry(Protocol):
|
||||||
|
|
||||||
async def initialize(self) -> None: ...
|
async def initialize(self) -> None: ...
|
||||||
|
|
||||||
async def get(self, identifier: str) -> List[RoutableObjectWithProvider]: ...
|
async def get(self, identifier: str) -> Optional[RoutableObjectWithProvider]: ...
|
||||||
|
|
||||||
def get_cached(self, identifier: str) -> List[RoutableObjectWithProvider]: ...
|
def get_cached(self, identifier: str) -> Optional[RoutableObjectWithProvider]: ...
|
||||||
|
|
||||||
|
async def update(
|
||||||
|
self, obj: RoutableObjectWithProvider
|
||||||
|
) -> RoutableObjectWithProvider: ...
|
||||||
|
|
||||||
# The current data structure allows multiple objects with the same identifier but different providers.
|
# The current data structure allows multiple objects with the same identifier but different providers.
|
||||||
# This is not ideal - we should have a single object that can be served by multiple providers,
|
# This is not ideal - we should have a single object that can be served by multiple providers,
|
||||||
|
@ -40,7 +44,7 @@ class DistributionRegistry(Protocol):
|
||||||
|
|
||||||
|
|
||||||
REGISTER_PREFIX = "distributions:registry"
|
REGISTER_PREFIX = "distributions:registry"
|
||||||
KEY_VERSION = "v1"
|
KEY_VERSION = "v2"
|
||||||
KEY_FORMAT = f"{REGISTER_PREFIX}:{KEY_VERSION}::" + "{type}:{identifier}"
|
KEY_FORMAT = f"{REGISTER_PREFIX}:{KEY_VERSION}::" + "{type}:{identifier}"
|
||||||
|
|
||||||
|
|
||||||
|
@ -54,19 +58,11 @@ def _parse_registry_values(values: List[str]) -> List[RoutableObjectWithProvider
|
||||||
"""Utility function to parse registry values into RoutableObjectWithProvider objects."""
|
"""Utility function to parse registry values into RoutableObjectWithProvider objects."""
|
||||||
all_objects = []
|
all_objects = []
|
||||||
for value in values:
|
for value in values:
|
||||||
try:
|
obj = pydantic.parse_obj_as(
|
||||||
objects_data = json.loads(value)
|
RoutableObjectWithProvider,
|
||||||
objects = [
|
json.loads(value),
|
||||||
pydantic.parse_obj_as(
|
)
|
||||||
RoutableObjectWithProvider,
|
all_objects.append(obj)
|
||||||
json.loads(obj_str),
|
|
||||||
)
|
|
||||||
for obj_str in objects_data
|
|
||||||
]
|
|
||||||
all_objects.extend(objects)
|
|
||||||
except Exception as e:
|
|
||||||
print(f"Error parsing value: {e}")
|
|
||||||
traceback.print_exc()
|
|
||||||
return all_objects
|
return all_objects
|
||||||
|
|
||||||
|
|
||||||
|
@ -79,46 +75,49 @@ class DiskDistributionRegistry(DistributionRegistry):
|
||||||
|
|
||||||
def get_cached(
|
def get_cached(
|
||||||
self, type: str, identifier: str
|
self, type: str, identifier: str
|
||||||
) -> List[RoutableObjectWithProvider]:
|
) -> Optional[RoutableObjectWithProvider]:
|
||||||
# Disk registry does not have a cache
|
# Disk registry does not have a cache
|
||||||
return []
|
raise NotImplementedError("Disk registry does not have a cache")
|
||||||
|
|
||||||
async def get_all(self) -> List[RoutableObjectWithProvider]:
|
async def get_all(self) -> List[RoutableObjectWithProvider]:
|
||||||
start_key, end_key = _get_registry_key_range()
|
start_key, end_key = _get_registry_key_range()
|
||||||
values = await self.kvstore.range(start_key, end_key)
|
values = await self.kvstore.range(start_key, end_key)
|
||||||
return _parse_registry_values(values)
|
return _parse_registry_values(values)
|
||||||
|
|
||||||
async def get(self, type: str, identifier: str) -> List[RoutableObjectWithProvider]:
|
async def get(
|
||||||
|
self, type: str, identifier: str
|
||||||
|
) -> Optional[RoutableObjectWithProvider]:
|
||||||
json_str = await self.kvstore.get(
|
json_str = await self.kvstore.get(
|
||||||
KEY_FORMAT.format(type=type, identifier=identifier)
|
KEY_FORMAT.format(type=type, identifier=identifier)
|
||||||
)
|
)
|
||||||
if not json_str:
|
if not json_str:
|
||||||
return []
|
return None
|
||||||
|
|
||||||
objects_data = json.loads(json_str)
|
objects_data = json.loads(json_str)
|
||||||
return [
|
# Return only the first object if any exist
|
||||||
pydantic.parse_obj_as(
|
if objects_data:
|
||||||
|
return pydantic.parse_obj_as(
|
||||||
RoutableObjectWithProvider,
|
RoutableObjectWithProvider,
|
||||||
json.loads(obj_str),
|
json.loads(objects_data),
|
||||||
)
|
)
|
||||||
for obj_str in objects_data
|
return None
|
||||||
]
|
|
||||||
|
|
||||||
async def register(self, obj: RoutableObjectWithProvider) -> bool:
|
async def update(self, obj: RoutableObjectWithProvider) -> None:
|
||||||
existing_objects = await self.get(obj.type, obj.identifier)
|
|
||||||
# dont register if the object's providerid already exists
|
|
||||||
for eobj in existing_objects:
|
|
||||||
if eobj.provider_id == obj.provider_id:
|
|
||||||
return False
|
|
||||||
|
|
||||||
existing_objects.append(obj)
|
|
||||||
|
|
||||||
objects_json = [
|
|
||||||
obj.model_dump_json() for obj in existing_objects
|
|
||||||
] # Fixed variable name
|
|
||||||
await self.kvstore.set(
|
await self.kvstore.set(
|
||||||
KEY_FORMAT.format(type=obj.type, identifier=obj.identifier),
|
KEY_FORMAT.format(type=obj.type, identifier=obj.identifier),
|
||||||
json.dumps(objects_json),
|
obj.model_dump_json(),
|
||||||
|
)
|
||||||
|
return obj
|
||||||
|
|
||||||
|
async def register(self, obj: RoutableObjectWithProvider) -> bool:
|
||||||
|
existing_obj = await self.get(obj.type, obj.identifier)
|
||||||
|
# dont register if the object's providerid already exists
|
||||||
|
if existing_obj and existing_obj.provider_id == obj.provider_id:
|
||||||
|
return False
|
||||||
|
|
||||||
|
await self.kvstore.set(
|
||||||
|
KEY_FORMAT.format(type=obj.type, identifier=obj.identifier),
|
||||||
|
obj.model_dump_json(),
|
||||||
)
|
)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@ -129,7 +128,7 @@ class DiskDistributionRegistry(DistributionRegistry):
|
||||||
class CachedDiskDistributionRegistry(DiskDistributionRegistry):
|
class CachedDiskDistributionRegistry(DiskDistributionRegistry):
|
||||||
def __init__(self, kvstore: KVStore):
|
def __init__(self, kvstore: KVStore):
|
||||||
super().__init__(kvstore)
|
super().__init__(kvstore)
|
||||||
self.cache: Dict[Tuple[str, str], List[RoutableObjectWithProvider]] = {}
|
self.cache: Dict[Tuple[str, str], RoutableObjectWithProvider] = {}
|
||||||
self._initialized = False
|
self._initialized = False
|
||||||
self._initialize_lock = asyncio.Lock()
|
self._initialize_lock = asyncio.Lock()
|
||||||
self._cache_lock = asyncio.Lock()
|
self._cache_lock = asyncio.Lock()
|
||||||
|
@ -156,13 +155,7 @@ class CachedDiskDistributionRegistry(DiskDistributionRegistry):
|
||||||
async with self._locked_cache() as cache:
|
async with self._locked_cache() as cache:
|
||||||
for obj in objects:
|
for obj in objects:
|
||||||
cache_key = (obj.type, obj.identifier)
|
cache_key = (obj.type, obj.identifier)
|
||||||
if cache_key not in cache:
|
cache[cache_key] = obj
|
||||||
cache[cache_key] = []
|
|
||||||
if not any(
|
|
||||||
cached_obj.provider_id == obj.provider_id
|
|
||||||
for cached_obj in cache[cache_key]
|
|
||||||
):
|
|
||||||
cache[cache_key].append(obj)
|
|
||||||
|
|
||||||
self._initialized = True
|
self._initialized = True
|
||||||
|
|
||||||
|
@ -171,28 +164,22 @@ class CachedDiskDistributionRegistry(DiskDistributionRegistry):
|
||||||
|
|
||||||
def get_cached(
|
def get_cached(
|
||||||
self, type: str, identifier: str
|
self, type: str, identifier: str
|
||||||
) -> List[RoutableObjectWithProvider]:
|
) -> Optional[RoutableObjectWithProvider]:
|
||||||
return self.cache.get((type, identifier), [])[:] # Return a copy
|
return self.cache.get((type, identifier), None)
|
||||||
|
|
||||||
async def get_all(self) -> List[RoutableObjectWithProvider]:
|
async def get_all(self) -> List[RoutableObjectWithProvider]:
|
||||||
await self._ensure_initialized()
|
await self._ensure_initialized()
|
||||||
async with self._locked_cache() as cache:
|
async with self._locked_cache() as cache:
|
||||||
return [item for sublist in cache.values() for item in sublist]
|
return list(cache.values())
|
||||||
|
|
||||||
async def get(self, type: str, identifier: str) -> List[RoutableObjectWithProvider]:
|
async def get(
|
||||||
|
self, type: str, identifier: str
|
||||||
|
) -> Optional[RoutableObjectWithProvider]:
|
||||||
await self._ensure_initialized()
|
await self._ensure_initialized()
|
||||||
cache_key = (type, identifier)
|
cache_key = (type, identifier)
|
||||||
|
|
||||||
async with self._locked_cache() as cache:
|
async with self._locked_cache() as cache:
|
||||||
if cache_key in cache:
|
return cache.get(cache_key, None)
|
||||||
return cache[cache_key][:]
|
|
||||||
|
|
||||||
objects = await super().get(type, identifier)
|
|
||||||
if objects:
|
|
||||||
async with self._locked_cache() as cache:
|
|
||||||
cache[cache_key] = objects
|
|
||||||
|
|
||||||
return objects
|
|
||||||
|
|
||||||
async def register(self, obj: RoutableObjectWithProvider) -> bool:
|
async def register(self, obj: RoutableObjectWithProvider) -> bool:
|
||||||
await self._ensure_initialized()
|
await self._ensure_initialized()
|
||||||
|
@ -201,16 +188,17 @@ class CachedDiskDistributionRegistry(DiskDistributionRegistry):
|
||||||
if success:
|
if success:
|
||||||
cache_key = (obj.type, obj.identifier)
|
cache_key = (obj.type, obj.identifier)
|
||||||
async with self._locked_cache() as cache:
|
async with self._locked_cache() as cache:
|
||||||
if cache_key not in cache:
|
cache[cache_key] = obj
|
||||||
cache[cache_key] = []
|
|
||||||
if not any(
|
|
||||||
cached_obj.provider_id == obj.provider_id
|
|
||||||
for cached_obj in cache[cache_key]
|
|
||||||
):
|
|
||||||
cache[cache_key].append(obj)
|
|
||||||
|
|
||||||
return success
|
return success
|
||||||
|
|
||||||
|
async def update(self, obj: RoutableObjectWithProvider) -> None:
|
||||||
|
await super().update(obj)
|
||||||
|
cache_key = (obj.type, obj.identifier)
|
||||||
|
async with self._locked_cache() as cache:
|
||||||
|
cache[cache_key] = obj
|
||||||
|
return obj
|
||||||
|
|
||||||
async def delete(self, type: str, identifier: str) -> None:
|
async def delete(self, type: str, identifier: str) -> None:
|
||||||
await super().delete(type, identifier)
|
await super().delete(type, identifier)
|
||||||
cache_key = (type, identifier)
|
cache_key = (type, identifier)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue