mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-16 21:09:27 +00:00
feat: support filters in file search
# What does this PR do? ## Test Plan
This commit is contained in:
parent
90d03552d4
commit
1c56b9133a
13 changed files with 449 additions and 63 deletions
|
|
@ -14,6 +14,7 @@ from llama_stack.apis.models import ModelType
|
|||
from llama_stack.apis.vector_io import (
|
||||
Chunk,
|
||||
QueryChunksResponse,
|
||||
SearchRankingOptions,
|
||||
VectorIO,
|
||||
VectorStoreDeleteResponse,
|
||||
VectorStoreListResponse,
|
||||
|
|
@ -246,7 +247,7 @@ class VectorIORouter(VectorIO):
|
|||
query: str | list[str],
|
||||
filters: dict[str, Any] | None = None,
|
||||
max_num_results: int | None = 10,
|
||||
ranking_options: dict[str, Any] | None = None,
|
||||
ranking_options: SearchRankingOptions | None = None,
|
||||
rewrite_query: bool | None = False,
|
||||
) -> VectorStoreSearchResponsePage:
|
||||
logger.debug(f"VectorIORouter.openai_search_vector_store: {vector_store_id}")
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue