memory routers working

This commit is contained in:
Xi Yan 2024-09-21 16:40:23 -07:00
parent 04f480d70c
commit f0580251a3
4 changed files with 97 additions and 39 deletions

View file

@ -4,7 +4,7 @@
# This source code is licensed under the terms described in the LICENSE file in # This source code is licensed under the terms described in the LICENSE file in
# the root directory of this source tree. # the root directory of this source tree.
from typing import Any, Dict, List, Tuple from typing import Any, AsyncGenerator, Dict, List, Tuple
from llama_stack.distribution.datatypes import Api from llama_stack.distribution.datatypes import Api
@ -24,13 +24,22 @@ class MemoryRouter(Memory):
self, self,
routing_table: RoutingTable, routing_table: RoutingTable,
) -> None: ) -> None:
self.api = Api.memory.value
self.routing_table = routing_table self.routing_table = routing_table
self.bank_id_to_type = {}
async def initialize(self) -> None: async def initialize(self) -> None:
pass await self.routing_table.initialize(self.api)
async def shutdown(self) -> None: async def shutdown(self) -> None:
pass await self.routing_table.shutdown(self.api)
def get_provider_from_bank_id(self, bank_id: str) -> Any:
bank_type = self.bank_id_to_type.get(bank_id)
if not bank_type:
raise ValueError(f"Could not find bank type for {bank_id}")
return self.routing_table.get_provider_impl(self.api, bank_type)
async def create_memory_bank( async def create_memory_bank(
self, self,
@ -39,9 +48,16 @@ class MemoryRouter(Memory):
url: Optional[URL] = None, url: Optional[URL] = None,
) -> MemoryBank: ) -> MemoryBank:
print("MemoryRouter: create_memory_bank") print("MemoryRouter: create_memory_bank")
bank_type = config.type
bank = await self.routing_table.get_provider_impl(
self.api, bank_type
).create_memory_bank(name, config, url)
self.bank_id_to_type[bank.bank_id] = bank_type
return bank
async def get_memory_bank(self, bank_id: str) -> Optional[MemoryBank]: async def get_memory_bank(self, bank_id: str) -> Optional[MemoryBank]:
print("MemoryRouter: get_memory_bank") print("MemoryRouter: get_memory_bank")
return await self.get_provider_from_bank_id(bank_id).get_memory_bank(bank_id)
async def insert_documents( async def insert_documents(
self, self,
@ -50,6 +66,9 @@ class MemoryRouter(Memory):
ttl_seconds: Optional[int] = None, ttl_seconds: Optional[int] = None,
) -> None: ) -> None:
print("MemoryRouter: insert_documents") print("MemoryRouter: insert_documents")
return await self.get_provider_from_bank_id(bank_id).insert_documents(
bank_id, documents, ttl_seconds
)
async def query_documents( async def query_documents(
self, self,
@ -57,7 +76,9 @@ class MemoryRouter(Memory):
query: InterleavedTextMedia, query: InterleavedTextMedia,
params: Optional[Dict[str, Any]] = None, params: Optional[Dict[str, Any]] = None,
) -> QueryDocumentsResponse: ) -> QueryDocumentsResponse:
print("query_documents") return await self.get_provider_from_bank_id(bank_id).query_documents(
bank_id, query, params
)
class InferenceRouter(Inference): class InferenceRouter(Inference):
@ -81,14 +102,13 @@ class InferenceRouter(Inference):
model: str, model: str,
messages: List[Message], messages: List[Message],
sampling_params: Optional[SamplingParams] = SamplingParams(), sampling_params: Optional[SamplingParams] = SamplingParams(),
# zero-shot tool definitions as input to the model tools: Optional[List[ToolDefinition]] = [],
tools: Optional[List[ToolDefinition]] = list,
tool_choice: Optional[ToolChoice] = ToolChoice.auto, tool_choice: Optional[ToolChoice] = ToolChoice.auto,
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json, tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
stream: Optional[bool] = False, stream: Optional[bool] = False,
logprobs: Optional[LogProbConfig] = None, logprobs: Optional[LogProbConfig] = None,
) -> AsyncGenerator: ) -> AsyncGenerator:
# TODO: we need to fix streaming response to align provider implementations with Protocol # TODO: we need to fix streaming response to align provider implementations with Protocol.
async for chunk in self.routing_table.get_provider_impl( async for chunk in self.routing_table.get_provider_impl(
self.api, model self.api, model
).chat_completion( ).chat_completion(

View file

@ -3,40 +3,40 @@ image_name: local
docker_image: null docker_image: null
conda_env: local conda_env: local
apis_to_serve: apis_to_serve:
- inference # - inference
# - memory - memory
- telemetry - telemetry
provider_map: provider_map:
telemetry: telemetry:
provider_id: meta-reference provider_id: meta-reference
config: {} config: {}
provider_routing_table: provider_routing_table:
inference: # inference:
- routing_key: Meta-Llama3.1-8B-Instruct # - routing_key: Meta-Llama3.1-8B-Instruct
provider_id: meta-reference
config:
model: Meta-Llama3.1-8B-Instruct
quantization: null
torch_seed: null
max_seq_len: 4096
max_batch_size: 1
# - routing_key: Meta-Llama3.1-8B
# provider_id: meta-reference
# config:
# model: Meta-Llama3.1-8B
# quantization: null
# torch_seed: null
# max_seq_len: 4096
# max_batch_size: 1
# memory:
# - routing_key: keyvalue
# provider_id: remote::pgvector
# config:
# host: localhost
# port: 5432
# db: vectordb
# user: vectoruser
# password: xxxx
# - routing_key: vector
# provider_id: meta-reference # provider_id: meta-reference
# config: {} # config:
# model: Meta-Llama3.1-8B-Instruct
# quantization: null
# torch_seed: null
# max_seq_len: 4096
# max_batch_size: 1
# - routing_key: Meta-Llama3.1-8B
# provider_id: meta-reference
# config:
# model: Meta-Llama3.1-8B
# quantization: null
# torch_seed: null
# max_seq_len: 4096
# max_batch_size: 1
memory:
- routing_key: keyvalue
provider_id: remote::pgvector
config:
host: localhost
port: 5432
db: vectordb
user: vectoruser
password: xxxx
- routing_key: vector
provider_id: meta-reference
config: {}

View file

@ -0,0 +1,38 @@
built_at: '2024-09-19T22:50:36.239761'
image_name: simple-local
docker_image: null
conda_env: simple-local
apis_to_serve:
- inference
- safety
- agents
- memory
provider_map:
inference:
provider_id: meta-reference
config:
model: Meta-Llama3.1-8B-Instruct
quantization: null
torch_seed: null
max_seq_len: 4096
max_batch_size: 1
safety:
provider_id: meta-reference
config:
llama_guard_shield:
model: Llama-Guard-3-8B
excluded_categories: []
disable_input_check: false
disable_output_check: false
prompt_guard_shield:
model: Prompt-Guard-86M
agents:
provider_id: meta-reference
config: {}
memory:
provider_id: meta-reference
config: {}
telemetry:
provider_id: meta-reference
config: {}
provider_routing_table: {}

View file

@ -57,7 +57,7 @@ class MetaReferenceInferenceImpl(Inference):
model: str, model: str,
messages: List[Message], messages: List[Message],
sampling_params: Optional[SamplingParams] = SamplingParams(), sampling_params: Optional[SamplingParams] = SamplingParams(),
tools: Optional[List[ToolDefinition]] = None, tools: Optional[List[ToolDefinition]] = [],
tool_choice: Optional[ToolChoice] = ToolChoice.auto, tool_choice: Optional[ToolChoice] = ToolChoice.auto,
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json, tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
stream: Optional[bool] = False, stream: Optional[bool] = False,
@ -70,7 +70,7 @@ class MetaReferenceInferenceImpl(Inference):
model=model, model=model,
messages=messages, messages=messages,
sampling_params=sampling_params, sampling_params=sampling_params,
tools=tools or [], tools=tools,
tool_choice=tool_choice, tool_choice=tool_choice,
tool_prompt_format=tool_prompt_format, tool_prompt_format=tool_prompt_format,
stream=stream, stream=stream,