mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-31 15:00:08 +00:00
Switched to using the existing Chunk type.
This commit is contained in:
parent
c3515530bb
commit
224d2d2891
2 changed files with 5 additions and 10 deletions
|
|
@ -8,8 +8,9 @@ from typing import Any, Dict, List, Optional, Protocol, runtime_checkable
|
|||
|
||||
from pydantic import BaseModel
|
||||
|
||||
from llama_stack.apis.common.content_types import URL
|
||||
from llama_stack.apis.common.content_types import URL, InterleavedContent
|
||||
from llama_stack.apis.preprocessing.preprocessors import Preprocessor
|
||||
from llama_stack.apis.vector_io import Chunk
|
||||
from llama_stack.schema_utils import json_schema_type, webmethod
|
||||
|
||||
|
||||
|
|
@ -28,16 +29,10 @@ class PreprocessingInput(BaseModel):
|
|||
PreprocessorOptions = Dict[str, Any]
|
||||
|
||||
|
||||
@json_schema_type
|
||||
class PreprocessingResult(BaseModel):
|
||||
metadata: dict[str, Any]
|
||||
data: str
|
||||
|
||||
|
||||
@json_schema_type
|
||||
class PreprocessingResponse(BaseModel):
|
||||
status: bool
|
||||
results: Optional[List[str | PreprocessingResult]]
|
||||
results: Optional[List[str | InterleavedContent | Chunk]]
|
||||
|
||||
|
||||
class PreprocessorStore(Protocol):
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue