mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
Bug fixes for memory
This commit is contained in:
parent
8cdc2f0cfb
commit
9ab27e852b
2 changed files with 5 additions and 5 deletions
|
@ -13,12 +13,12 @@ from typing import Any, Dict, List, Optional
|
|||
|
||||
import fire
|
||||
import httpx
|
||||
|
||||
from llama_stack.distribution.datatypes import RemoteProviderConfig
|
||||
from termcolor import cprint
|
||||
|
||||
from .memory import * # noqa: F403
|
||||
from .common.file_utils import data_url_from_file
|
||||
from llama_stack.distribution.datatypes import RemoteProviderConfig
|
||||
|
||||
from llama_stack.apis.memory import * # noqa: F403
|
||||
from llama_stack.providers.utils.memory.file_utils import data_url_from_file
|
||||
|
||||
|
||||
async def get_client_impl(config: RemoteProviderConfig, _deps: Any) -> Memory:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue