mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
memory routers working
This commit is contained in:
parent
04f480d70c
commit
f0580251a3
4 changed files with 97 additions and 39 deletions
|
@ -4,7 +4,7 @@
|
||||||
# This source code is licensed under the terms described in the LICENSE file in
|
# This source code is licensed under the terms described in the LICENSE file in
|
||||||
# the root directory of this source tree.
|
# the root directory of this source tree.
|
||||||
|
|
||||||
from typing import Any, Dict, List, Tuple
|
from typing import Any, AsyncGenerator, Dict, List, Tuple
|
||||||
|
|
||||||
from llama_stack.distribution.datatypes import Api
|
from llama_stack.distribution.datatypes import Api
|
||||||
|
|
||||||
|
@ -24,13 +24,22 @@ class MemoryRouter(Memory):
|
||||||
self,
|
self,
|
||||||
routing_table: RoutingTable,
|
routing_table: RoutingTable,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
self.api = Api.memory.value
|
||||||
self.routing_table = routing_table
|
self.routing_table = routing_table
|
||||||
|
self.bank_id_to_type = {}
|
||||||
|
|
||||||
async def initialize(self) -> None:
|
async def initialize(self) -> None:
|
||||||
pass
|
await self.routing_table.initialize(self.api)
|
||||||
|
|
||||||
async def shutdown(self) -> None:
|
async def shutdown(self) -> None:
|
||||||
pass
|
await self.routing_table.shutdown(self.api)
|
||||||
|
|
||||||
|
def get_provider_from_bank_id(self, bank_id: str) -> Any:
|
||||||
|
bank_type = self.bank_id_to_type.get(bank_id)
|
||||||
|
if not bank_type:
|
||||||
|
raise ValueError(f"Could not find bank type for {bank_id}")
|
||||||
|
|
||||||
|
return self.routing_table.get_provider_impl(self.api, bank_type)
|
||||||
|
|
||||||
async def create_memory_bank(
|
async def create_memory_bank(
|
||||||
self,
|
self,
|
||||||
|
@ -39,9 +48,16 @@ class MemoryRouter(Memory):
|
||||||
url: Optional[URL] = None,
|
url: Optional[URL] = None,
|
||||||
) -> MemoryBank:
|
) -> MemoryBank:
|
||||||
print("MemoryRouter: create_memory_bank")
|
print("MemoryRouter: create_memory_bank")
|
||||||
|
bank_type = config.type
|
||||||
|
bank = await self.routing_table.get_provider_impl(
|
||||||
|
self.api, bank_type
|
||||||
|
).create_memory_bank(name, config, url)
|
||||||
|
self.bank_id_to_type[bank.bank_id] = bank_type
|
||||||
|
return bank
|
||||||
|
|
||||||
async def get_memory_bank(self, bank_id: str) -> Optional[MemoryBank]:
|
async def get_memory_bank(self, bank_id: str) -> Optional[MemoryBank]:
|
||||||
print("MemoryRouter: get_memory_bank")
|
print("MemoryRouter: get_memory_bank")
|
||||||
|
return await self.get_provider_from_bank_id(bank_id).get_memory_bank(bank_id)
|
||||||
|
|
||||||
async def insert_documents(
|
async def insert_documents(
|
||||||
self,
|
self,
|
||||||
|
@ -50,6 +66,9 @@ class MemoryRouter(Memory):
|
||||||
ttl_seconds: Optional[int] = None,
|
ttl_seconds: Optional[int] = None,
|
||||||
) -> None:
|
) -> None:
|
||||||
print("MemoryRouter: insert_documents")
|
print("MemoryRouter: insert_documents")
|
||||||
|
return await self.get_provider_from_bank_id(bank_id).insert_documents(
|
||||||
|
bank_id, documents, ttl_seconds
|
||||||
|
)
|
||||||
|
|
||||||
async def query_documents(
|
async def query_documents(
|
||||||
self,
|
self,
|
||||||
|
@ -57,7 +76,9 @@ class MemoryRouter(Memory):
|
||||||
query: InterleavedTextMedia,
|
query: InterleavedTextMedia,
|
||||||
params: Optional[Dict[str, Any]] = None,
|
params: Optional[Dict[str, Any]] = None,
|
||||||
) -> QueryDocumentsResponse:
|
) -> QueryDocumentsResponse:
|
||||||
print("query_documents")
|
return await self.get_provider_from_bank_id(bank_id).query_documents(
|
||||||
|
bank_id, query, params
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class InferenceRouter(Inference):
|
class InferenceRouter(Inference):
|
||||||
|
@ -81,14 +102,13 @@ class InferenceRouter(Inference):
|
||||||
model: str,
|
model: str,
|
||||||
messages: List[Message],
|
messages: List[Message],
|
||||||
sampling_params: Optional[SamplingParams] = SamplingParams(),
|
sampling_params: Optional[SamplingParams] = SamplingParams(),
|
||||||
# zero-shot tool definitions as input to the model
|
tools: Optional[List[ToolDefinition]] = [],
|
||||||
tools: Optional[List[ToolDefinition]] = list,
|
|
||||||
tool_choice: Optional[ToolChoice] = ToolChoice.auto,
|
tool_choice: Optional[ToolChoice] = ToolChoice.auto,
|
||||||
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
|
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
|
||||||
stream: Optional[bool] = False,
|
stream: Optional[bool] = False,
|
||||||
logprobs: Optional[LogProbConfig] = None,
|
logprobs: Optional[LogProbConfig] = None,
|
||||||
) -> AsyncGenerator:
|
) -> AsyncGenerator:
|
||||||
# TODO: we need to fix streaming response to align provider implementations with Protocol
|
# TODO: we need to fix streaming response to align provider implementations with Protocol.
|
||||||
async for chunk in self.routing_table.get_provider_impl(
|
async for chunk in self.routing_table.get_provider_impl(
|
||||||
self.api, model
|
self.api, model
|
||||||
).chat_completion(
|
).chat_completion(
|
||||||
|
|
|
@ -3,40 +3,40 @@ image_name: local
|
||||||
docker_image: null
|
docker_image: null
|
||||||
conda_env: local
|
conda_env: local
|
||||||
apis_to_serve:
|
apis_to_serve:
|
||||||
- inference
|
# - inference
|
||||||
# - memory
|
- memory
|
||||||
- telemetry
|
- telemetry
|
||||||
provider_map:
|
provider_map:
|
||||||
telemetry:
|
telemetry:
|
||||||
provider_id: meta-reference
|
provider_id: meta-reference
|
||||||
config: {}
|
config: {}
|
||||||
provider_routing_table:
|
provider_routing_table:
|
||||||
inference:
|
# inference:
|
||||||
- routing_key: Meta-Llama3.1-8B-Instruct
|
# - routing_key: Meta-Llama3.1-8B-Instruct
|
||||||
provider_id: meta-reference
|
|
||||||
config:
|
|
||||||
model: Meta-Llama3.1-8B-Instruct
|
|
||||||
quantization: null
|
|
||||||
torch_seed: null
|
|
||||||
max_seq_len: 4096
|
|
||||||
max_batch_size: 1
|
|
||||||
# - routing_key: Meta-Llama3.1-8B
|
|
||||||
# provider_id: meta-reference
|
|
||||||
# config:
|
|
||||||
# model: Meta-Llama3.1-8B
|
|
||||||
# quantization: null
|
|
||||||
# torch_seed: null
|
|
||||||
# max_seq_len: 4096
|
|
||||||
# max_batch_size: 1
|
|
||||||
# memory:
|
|
||||||
# - routing_key: keyvalue
|
|
||||||
# provider_id: remote::pgvector
|
|
||||||
# config:
|
|
||||||
# host: localhost
|
|
||||||
# port: 5432
|
|
||||||
# db: vectordb
|
|
||||||
# user: vectoruser
|
|
||||||
# password: xxxx
|
|
||||||
# - routing_key: vector
|
|
||||||
# provider_id: meta-reference
|
# provider_id: meta-reference
|
||||||
# config: {}
|
# config:
|
||||||
|
# model: Meta-Llama3.1-8B-Instruct
|
||||||
|
# quantization: null
|
||||||
|
# torch_seed: null
|
||||||
|
# max_seq_len: 4096
|
||||||
|
# max_batch_size: 1
|
||||||
|
# - routing_key: Meta-Llama3.1-8B
|
||||||
|
# provider_id: meta-reference
|
||||||
|
# config:
|
||||||
|
# model: Meta-Llama3.1-8B
|
||||||
|
# quantization: null
|
||||||
|
# torch_seed: null
|
||||||
|
# max_seq_len: 4096
|
||||||
|
# max_batch_size: 1
|
||||||
|
memory:
|
||||||
|
- routing_key: keyvalue
|
||||||
|
provider_id: remote::pgvector
|
||||||
|
config:
|
||||||
|
host: localhost
|
||||||
|
port: 5432
|
||||||
|
db: vectordb
|
||||||
|
user: vectoruser
|
||||||
|
password: xxxx
|
||||||
|
- routing_key: vector
|
||||||
|
provider_id: meta-reference
|
||||||
|
config: {}
|
||||||
|
|
38
llama_stack/examples/simple-local-run.yaml
Normal file
38
llama_stack/examples/simple-local-run.yaml
Normal file
|
@ -0,0 +1,38 @@
|
||||||
|
built_at: '2024-09-19T22:50:36.239761'
|
||||||
|
image_name: simple-local
|
||||||
|
docker_image: null
|
||||||
|
conda_env: simple-local
|
||||||
|
apis_to_serve:
|
||||||
|
- inference
|
||||||
|
- safety
|
||||||
|
- agents
|
||||||
|
- memory
|
||||||
|
provider_map:
|
||||||
|
inference:
|
||||||
|
provider_id: meta-reference
|
||||||
|
config:
|
||||||
|
model: Meta-Llama3.1-8B-Instruct
|
||||||
|
quantization: null
|
||||||
|
torch_seed: null
|
||||||
|
max_seq_len: 4096
|
||||||
|
max_batch_size: 1
|
||||||
|
safety:
|
||||||
|
provider_id: meta-reference
|
||||||
|
config:
|
||||||
|
llama_guard_shield:
|
||||||
|
model: Llama-Guard-3-8B
|
||||||
|
excluded_categories: []
|
||||||
|
disable_input_check: false
|
||||||
|
disable_output_check: false
|
||||||
|
prompt_guard_shield:
|
||||||
|
model: Prompt-Guard-86M
|
||||||
|
agents:
|
||||||
|
provider_id: meta-reference
|
||||||
|
config: {}
|
||||||
|
memory:
|
||||||
|
provider_id: meta-reference
|
||||||
|
config: {}
|
||||||
|
telemetry:
|
||||||
|
provider_id: meta-reference
|
||||||
|
config: {}
|
||||||
|
provider_routing_table: {}
|
|
@ -57,7 +57,7 @@ class MetaReferenceInferenceImpl(Inference):
|
||||||
model: str,
|
model: str,
|
||||||
messages: List[Message],
|
messages: List[Message],
|
||||||
sampling_params: Optional[SamplingParams] = SamplingParams(),
|
sampling_params: Optional[SamplingParams] = SamplingParams(),
|
||||||
tools: Optional[List[ToolDefinition]] = None,
|
tools: Optional[List[ToolDefinition]] = [],
|
||||||
tool_choice: Optional[ToolChoice] = ToolChoice.auto,
|
tool_choice: Optional[ToolChoice] = ToolChoice.auto,
|
||||||
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
|
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
|
||||||
stream: Optional[bool] = False,
|
stream: Optional[bool] = False,
|
||||||
|
@ -70,7 +70,7 @@ class MetaReferenceInferenceImpl(Inference):
|
||||||
model=model,
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
sampling_params=sampling_params,
|
sampling_params=sampling_params,
|
||||||
tools=tools or [],
|
tools=tools,
|
||||||
tool_choice=tool_choice,
|
tool_choice=tool_choice,
|
||||||
tool_prompt_format=tool_prompt_format,
|
tool_prompt_format=tool_prompt_format,
|
||||||
stream=stream,
|
stream=stream,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue