diff --git a/llama_stack/apis/vector_io/vector_io.py b/llama_stack/apis/vector_io/vector_io.py index e07175c49..f688be8a4 100644 --- a/llama_stack/apis/vector_io/vector_io.py +++ b/llama_stack/apis/vector_io/vector_io.py @@ -11,8 +11,6 @@ import uuid from typing import Annotated, Any, Literal, Protocol, runtime_checkable -from pydantic import BaseModel, Field - from llama_stack.apis.inference import InterleavedContent from llama_stack.apis.vector_dbs import VectorDB from llama_stack.apis.version import LLAMA_STACK_API_V1 @@ -21,6 +19,8 @@ from llama_stack.providers.utils.vector_io.vector_utils import generate_chunk_id from llama_stack.schema_utils import json_schema_type, webmethod from llama_stack.strong_typing.schema import register_schema +from pydantic import BaseModel, Field + @json_schema_type class ChunkMetadata(BaseModel): @@ -350,7 +350,12 @@ class VectorStoreFileLastError(BaseModel): message: str -VectorStoreFileStatus = Literal["completed"] | Literal["in_progress"] | Literal["cancelled"] | Literal["failed"] +VectorStoreFileStatus = ( + Literal["completed"] + | Literal["in_progress"] + | Literal["cancelled"] + | Literal["failed"] +) register_schema(VectorStoreFileStatus, name="VectorStoreFileStatus") @@ -556,7 +561,9 @@ class VectorIO(Protocol): """ ... - @webmethod(route="/vector_stores/{vector_store_id}", method="GET", level=LLAMA_STACK_API_V1) + @webmethod( + route="/vector_stores/{vector_store_id}", method="GET", level=LLAMA_STACK_API_V1 + ) async def openai_retrieve_vector_store( self, vector_store_id: str,