mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
Still retrieve the file_response in openai_vector_store_mixin
This is needed to get the filename of our file, even though we don't need its actual contents here anymore. Signed-off-by: Ben Browning <bbrownin@redhat.com>
This commit is contained in:
parent
8bf1d91d38
commit
0f7d487dca
1 changed files with 1 additions and 0 deletions
|
@ -543,6 +543,7 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
chunk_overlap_tokens = 400
|
||||
|
||||
try:
|
||||
file_response = await self.files_api.openai_retrieve_file(file_id)
|
||||
tool_result = await self.tool_runtime_api.invoke_tool(
|
||||
"convert_file_to_text",
|
||||
{"file_id": file_id},
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue