From 73c18feac48ead2293c3dc7f16707a0ac2ae2603 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Tue, 24 Jun 2025 18:55:56 +0530 Subject: [PATCH] fix: update the signature of openai_list_files_in_vector_store in all VectorIO impls (#2503) --- .../providers/remote/vector_io/chroma/chroma.py | 16 ++++++++++------ .../providers/remote/vector_io/milvus/milvus.py | 16 ++++++++++------ .../providers/remote/vector_io/qdrant/qdrant.py | 16 ++++++++++------ .../utils/memory/openai_vector_store_mixin.py | 16 +++++++--------- 4 files changed, 37 insertions(+), 27 deletions(-) diff --git a/llama_stack/providers/remote/vector_io/chroma/chroma.py b/llama_stack/providers/remote/vector_io/chroma/chroma.py index 9f206aaee..407bdda56 100644 --- a/llama_stack/providers/remote/vector_io/chroma/chroma.py +++ b/llama_stack/providers/remote/vector_io/chroma/chroma.py @@ -19,17 +19,16 @@ from llama_stack.apis.vector_io import ( QueryChunksResponse, SearchRankingOptions, VectorIO, + VectorStoreChunkingStrategy, VectorStoreDeleteResponse, + VectorStoreFileContentsResponse, + VectorStoreFileObject, + VectorStoreFileStatus, + VectorStoreListFilesResponse, VectorStoreListResponse, VectorStoreObject, VectorStoreSearchResponsePage, ) -from llama_stack.apis.vector_io.vector_io import ( - VectorStoreChunkingStrategy, - VectorStoreFileContentsResponse, - VectorStoreFileObject, - VectorStoreListFilesResponse, -) from llama_stack.providers.datatypes import Api, VectorDBsProtocolPrivate from llama_stack.providers.inline.vector_io.chroma import ChromaVectorIOConfig as InlineChromaVectorIOConfig from llama_stack.providers.utils.memory.vector_store import ( @@ -272,6 +271,11 @@ class ChromaVectorIOAdapter(VectorIO, VectorDBsProtocolPrivate): async def openai_list_files_in_vector_store( self, vector_store_id: str, + limit: int | None = 20, + order: str | None = "desc", + after: str | None = None, + before: str | None = None, + filter: VectorStoreFileStatus | None = None, ) -> VectorStoreListFilesResponse: raise NotImplementedError("OpenAI Vector Stores API is not supported in Chroma") diff --git a/llama_stack/providers/remote/vector_io/milvus/milvus.py b/llama_stack/providers/remote/vector_io/milvus/milvus.py index 737a46bac..09d8520d1 100644 --- a/llama_stack/providers/remote/vector_io/milvus/milvus.py +++ b/llama_stack/providers/remote/vector_io/milvus/milvus.py @@ -21,17 +21,16 @@ from llama_stack.apis.vector_io import ( QueryChunksResponse, SearchRankingOptions, VectorIO, + VectorStoreChunkingStrategy, VectorStoreDeleteResponse, + VectorStoreFileContentsResponse, + VectorStoreFileObject, + VectorStoreFileStatus, + VectorStoreListFilesResponse, VectorStoreListResponse, VectorStoreObject, VectorStoreSearchResponsePage, ) -from llama_stack.apis.vector_io.vector_io import ( - VectorStoreChunkingStrategy, - VectorStoreFileContentsResponse, - VectorStoreFileObject, - VectorStoreListFilesResponse, -) from llama_stack.providers.datatypes import Api, VectorDBsProtocolPrivate from llama_stack.providers.inline.vector_io.milvus import MilvusVectorIOConfig as InlineMilvusVectorIOConfig from llama_stack.providers.utils.memory.vector_store import ( @@ -270,6 +269,11 @@ class MilvusVectorIOAdapter(VectorIO, VectorDBsProtocolPrivate): async def openai_list_files_in_vector_store( self, vector_store_id: str, + limit: int | None = 20, + order: str | None = "desc", + after: str | None = None, + before: str | None = None, + filter: VectorStoreFileStatus | None = None, ) -> VectorStoreListFilesResponse: raise NotImplementedError("OpenAI Vector Stores API is not supported in Milvus") diff --git a/llama_stack/providers/remote/vector_io/qdrant/qdrant.py b/llama_stack/providers/remote/vector_io/qdrant/qdrant.py index e00fdf84e..694a1bf93 100644 --- a/llama_stack/providers/remote/vector_io/qdrant/qdrant.py +++ b/llama_stack/providers/remote/vector_io/qdrant/qdrant.py @@ -19,17 +19,16 @@ from llama_stack.apis.vector_io import ( QueryChunksResponse, SearchRankingOptions, VectorIO, + VectorStoreChunkingStrategy, VectorStoreDeleteResponse, + VectorStoreFileContentsResponse, + VectorStoreFileObject, + VectorStoreFileStatus, + VectorStoreListFilesResponse, VectorStoreListResponse, VectorStoreObject, VectorStoreSearchResponsePage, ) -from llama_stack.apis.vector_io.vector_io import ( - VectorStoreChunkingStrategy, - VectorStoreFileContentsResponse, - VectorStoreFileObject, - VectorStoreListFilesResponse, -) from llama_stack.providers.datatypes import Api, VectorDBsProtocolPrivate from llama_stack.providers.inline.vector_io.qdrant import QdrantVectorIOConfig as InlineQdrantVectorIOConfig from llama_stack.providers.utils.memory.vector_store import ( @@ -272,6 +271,11 @@ class QdrantVectorIOAdapter(VectorIO, VectorDBsProtocolPrivate): async def openai_list_files_in_vector_store( self, vector_store_id: str, + limit: int | None = 20, + order: str | None = "desc", + after: str | None = None, + before: str | None = None, + filter: VectorStoreFileStatus | None = None, ) -> VectorStoreListFilesResponse: raise NotImplementedError("OpenAI Vector Stores API is not supported in Qdrant") diff --git a/llama_stack/providers/utils/memory/openai_vector_store_mixin.py b/llama_stack/providers/utils/memory/openai_vector_store_mixin.py index 9c0e1dbe7..c741f7045 100644 --- a/llama_stack/providers/utils/memory/openai_vector_store_mixin.py +++ b/llama_stack/providers/utils/memory/openai_vector_store_mixin.py @@ -16,20 +16,14 @@ from llama_stack.apis.files import Files from llama_stack.apis.files.files import OpenAIFileObject from llama_stack.apis.vector_dbs import VectorDB from llama_stack.apis.vector_io import ( + Chunk, QueryChunksResponse, SearchRankingOptions, - VectorStoreContent, - VectorStoreDeleteResponse, - VectorStoreListResponse, - VectorStoreObject, - VectorStoreSearchResponse, - VectorStoreSearchResponsePage, -) -from llama_stack.apis.vector_io.vector_io import ( - Chunk, VectorStoreChunkingStrategy, VectorStoreChunkingStrategyAuto, VectorStoreChunkingStrategyStatic, + VectorStoreContent, + VectorStoreDeleteResponse, VectorStoreFileContentsResponse, VectorStoreFileCounts, VectorStoreFileDeleteResponse, @@ -37,6 +31,10 @@ from llama_stack.apis.vector_io.vector_io import ( VectorStoreFileObject, VectorStoreFileStatus, VectorStoreListFilesResponse, + VectorStoreListResponse, + VectorStoreObject, + VectorStoreSearchResponse, + VectorStoreSearchResponsePage, ) from llama_stack.providers.utils.memory.vector_store import content_from_data_and_mime_type, make_overlapped_chunks