mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
Merge 1e5780b8dc
into d266c59c2a
This commit is contained in:
commit
f7cf325690
4 changed files with 44 additions and 5 deletions
|
@ -92,6 +92,22 @@ class Chunk(BaseModel):
|
||||||
|
|
||||||
return generate_chunk_id(str(uuid.uuid4()), str(self.content))
|
return generate_chunk_id(str(uuid.uuid4()), str(self.content))
|
||||||
|
|
||||||
|
@property
|
||||||
|
def document_id(self) -> str | None:
|
||||||
|
"""Returns the document_id from either metadata or chunk_metadata, with metadata taking precedence."""
|
||||||
|
# Check metadata first (takes precedence)
|
||||||
|
doc_id = self.metadata.get("document_id")
|
||||||
|
if isinstance(doc_id, str):
|
||||||
|
return doc_id
|
||||||
|
|
||||||
|
# Fall back to chunk_metadata if available
|
||||||
|
if self.chunk_metadata is not None:
|
||||||
|
chunk_doc_id = getattr(self.chunk_metadata, "document_id", None)
|
||||||
|
if isinstance(chunk_doc_id, str):
|
||||||
|
return chunk_doc_id
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
@json_schema_type
|
@json_schema_type
|
||||||
class QueryChunksResponse(BaseModel):
|
class QueryChunksResponse(BaseModel):
|
||||||
|
|
|
@ -5,6 +5,7 @@
|
||||||
# the root directory of this source tree.
|
# the root directory of this source tree.
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import logging # allow-direct-logging
|
||||||
import uuid
|
import uuid
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
|
@ -101,11 +102,14 @@ class VectorIORouter(VectorIO):
|
||||||
chunks: list[Chunk],
|
chunks: list[Chunk],
|
||||||
ttl_seconds: int | None = None,
|
ttl_seconds: int | None = None,
|
||||||
) -> None:
|
) -> None:
|
||||||
logger.debug(
|
if logger.isEnabledFor(logging.DEBUG):
|
||||||
f"VectorIORouter.insert_chunks: {vector_db_id}, {len(chunks)} chunks, ttl_seconds={ttl_seconds}, chunk_ids={[chunk.metadata['document_id'] for chunk in chunks[:3]]}{' and more...' if len(chunks) > 3 else ''}",
|
doc_ids = [chunk.document_id for chunk in chunks[:3]]
|
||||||
)
|
logger.debug(
|
||||||
|
f"VectorIORouter.insert_chunks: {vector_db_id}, {len(chunks)} chunks, "
|
||||||
|
f"ttl_seconds={ttl_seconds}, chunk_ids={doc_ids}{' and more...' if len(chunks) > 3 else ''}"
|
||||||
|
)
|
||||||
provider = await self.routing_table.get_provider_impl(vector_db_id)
|
provider = await self.routing_table.get_provider_impl(vector_db_id)
|
||||||
return await provider.insert_chunks(vector_db_id, chunks, ttl_seconds)
|
await provider.insert_chunks(vector_db_id, chunks, ttl_seconds)
|
||||||
|
|
||||||
async def query_chunks(
|
async def query_chunks(
|
||||||
self,
|
self,
|
||||||
|
|
|
@ -278,7 +278,7 @@ class MemoryToolRuntimeImpl(ToolGroupsProtocolPrivate, ToolRuntime, RAGToolRunti
|
||||||
return RAGQueryResult(
|
return RAGQueryResult(
|
||||||
content=picked,
|
content=picked,
|
||||||
metadata={
|
metadata={
|
||||||
"document_ids": [c.metadata["document_id"] for c in chunks[: len(picked)]],
|
"document_ids": [c.document_id for c in chunks[: len(picked)]],
|
||||||
"chunks": [c.content for c in chunks[: len(picked)]],
|
"chunks": [c.content for c in chunks[: len(picked)]],
|
||||||
"scores": scores[: len(picked)],
|
"scores": scores[: len(picked)],
|
||||||
"vector_db_ids": [c.metadata["vector_db_id"] for c in chunks[: len(picked)]],
|
"vector_db_ids": [c.metadata["vector_db_id"] for c in chunks[: len(picked)]],
|
||||||
|
|
|
@ -113,6 +113,25 @@ async def test_insert_chunks_missing_db_raises(vector_io_adapter):
|
||||||
await vector_io_adapter.insert_chunks("db_not_exist", [])
|
await vector_io_adapter.insert_chunks("db_not_exist", [])
|
||||||
|
|
||||||
|
|
||||||
|
async def test_insert_chunks_with_missing_document_id(vector_io_adapter):
|
||||||
|
"""Ensure no KeyError when document_id is missing or in different places."""
|
||||||
|
from llama_stack.apis.vector_io import Chunk, ChunkMetadata
|
||||||
|
|
||||||
|
fake_index = AsyncMock()
|
||||||
|
vector_io_adapter.cache["db1"] = fake_index
|
||||||
|
|
||||||
|
# Various document_id scenarios that shouldn't crash
|
||||||
|
chunks = [
|
||||||
|
Chunk(content="has doc_id in metadata", metadata={"document_id": "doc-1"}),
|
||||||
|
Chunk(content="no doc_id anywhere", metadata={"source": "test"}),
|
||||||
|
Chunk(content="doc_id in chunk_metadata", chunk_metadata=ChunkMetadata(document_id="doc-3")),
|
||||||
|
]
|
||||||
|
|
||||||
|
# Should work without KeyError
|
||||||
|
await vector_io_adapter.insert_chunks("db1", chunks)
|
||||||
|
fake_index.insert_chunks.assert_awaited_once()
|
||||||
|
|
||||||
|
|
||||||
async def test_query_chunks_calls_underlying_index_and_returns(vector_io_adapter):
|
async def test_query_chunks_calls_underlying_index_and_returns(vector_io_adapter):
|
||||||
expected = QueryChunksResponse(chunks=[Chunk(content="c1")], scores=[0.1])
|
expected = QueryChunksResponse(chunks=[Chunk(content="c1")], scores=[0.1])
|
||||||
fake_index = AsyncMock(query_chunks=AsyncMock(return_value=expected))
|
fake_index = AsyncMock(query_chunks=AsyncMock(return_value=expected))
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue