mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-31 21:44:32 +00:00
Added output type to PreprocessorResponse.
This commit is contained in:
parent
b981181b25
commit
4c81a72214
7 changed files with 33 additions and 25 deletions
|
|
@ -56,7 +56,8 @@ PreprocessorChain = List[PreprocessorChainElement]
|
|||
|
||||
@json_schema_type
|
||||
class PreprocessorResponse(BaseModel):
|
||||
status: bool
|
||||
success: bool
|
||||
preprocessor_output_type: PreprocessingDataType
|
||||
results: Optional[List[str | InterleavedContent | Chunk]] = None
|
||||
|
||||
|
||||
|
|
@ -84,5 +85,4 @@ class Preprocessing(Protocol):
|
|||
self,
|
||||
preprocessors: PreprocessorChain,
|
||||
preprocessor_inputs: List[PreprocessorInput],
|
||||
is_rag_chain: Optional[bool] = False,
|
||||
) -> PreprocessorResponse: ...
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue