mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-10 04:08:31 +00:00
Added lazy initialization to the docling provider.
This commit is contained in:
parent
275fdbc23f
commit
e895bb111c
1 changed files with 9 additions and 5 deletions
|
@ -36,13 +36,10 @@ class InclineDoclingPreprocessorImpl(Preprocessing, PreprocessorsProtocolPrivate
|
||||||
|
|
||||||
def __init__(self, config: InlineDoclingConfig) -> None:
|
def __init__(self, config: InlineDoclingConfig) -> None:
|
||||||
self.config = config
|
self.config = config
|
||||||
self.converter = DocumentConverter()
|
self.converter = None
|
||||||
self.chunker = None
|
self.chunker = None
|
||||||
|
|
||||||
async def initialize(self) -> None:
|
async def initialize(self) -> None: ...
|
||||||
if self.config.chunk:
|
|
||||||
# TODO: docling should use Llama Stack's inference API instead of handling tokenization by itself
|
|
||||||
self.chunker = HybridChunker()
|
|
||||||
|
|
||||||
async def shutdown(self) -> None: ...
|
async def shutdown(self) -> None: ...
|
||||||
|
|
||||||
|
@ -56,6 +53,13 @@ class InclineDoclingPreprocessorImpl(Preprocessing, PreprocessorsProtocolPrivate
|
||||||
preprocessor_inputs: List[PreprocessingDataElement],
|
preprocessor_inputs: List[PreprocessingDataElement],
|
||||||
options: Optional[PreprocessorOptions] = None,
|
options: Optional[PreprocessorOptions] = None,
|
||||||
) -> PreprocessorResponse:
|
) -> PreprocessorResponse:
|
||||||
|
if self.converter is None:
|
||||||
|
# this is the first time this method is called
|
||||||
|
self.converter = DocumentConverter()
|
||||||
|
if self.config.chunk and self.chunker is None:
|
||||||
|
# TODO: docling should use Llama Stack's inference API instead of handling tokenization by itself
|
||||||
|
self.chunker = HybridChunker()
|
||||||
|
|
||||||
results = []
|
results = []
|
||||||
|
|
||||||
for inp in preprocessor_inputs:
|
for inp in preprocessor_inputs:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue