mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-30 07:39:38 +00:00
add back wizard for build
This commit is contained in:
parent
055770a791
commit
b611e8bbd2
2 changed files with 16 additions and 7 deletions
|
@ -27,7 +27,7 @@ class StackBuild(Subcommand):
|
||||||
|
|
||||||
def _add_arguments(self):
|
def _add_arguments(self):
|
||||||
self.parser.add_argument(
|
self.parser.add_argument(
|
||||||
"config",
|
"--config",
|
||||||
type=str,
|
type=str,
|
||||||
help="Path to a config file to use for the build. You may find example configs in llama_stack/distribution/example_configs",
|
help="Path to a config file to use for the build. You may find example configs in llama_stack/distribution/example_configs",
|
||||||
)
|
)
|
||||||
|
@ -44,9 +44,10 @@ class StackBuild(Subcommand):
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
from llama_stack.distribution.build import ApiInput, build_image, ImageType
|
||||||
|
|
||||||
from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR
|
from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR
|
||||||
from llama_stack.distribution.utils.serialize import EnumEncoder
|
from llama_stack.distribution.utils.serialize import EnumEncoder
|
||||||
from llama_stack.distribution.build import ApiInput, build_image, ImageType
|
|
||||||
from termcolor import cprint
|
from termcolor import cprint
|
||||||
|
|
||||||
# save build.yaml spec for building same distribution again
|
# save build.yaml spec for building same distribution again
|
||||||
|
@ -74,13 +75,15 @@ class StackBuild(Subcommand):
|
||||||
)
|
)
|
||||||
|
|
||||||
def _run_stack_build_command(self, args: argparse.Namespace) -> None:
|
def _run_stack_build_command(self, args: argparse.Namespace) -> None:
|
||||||
from llama_stack.distribution.utils.prompt_for_config import prompt_for_config
|
|
||||||
from llama_stack.distribution.utils.dynamic import instantiate_class_type
|
from llama_stack.distribution.utils.dynamic import instantiate_class_type
|
||||||
|
from llama_stack.distribution.utils.prompt_for_config import prompt_for_config
|
||||||
|
|
||||||
if not args.config:
|
if not args.config:
|
||||||
self.parser.error(
|
# self.parser.error(
|
||||||
"No config file specified. Please use `llama stack build /path/to/*-build.yaml`. Example config files can be found in llama_stack/distribution/example_configs"
|
# "No config file specified. Please use `llama stack build /path/to/*-build.yaml`. Example config files can be found in llama_stack/distribution/example_configs"
|
||||||
)
|
# )
|
||||||
|
build_config = prompt_for_config(BuildConfig, None)
|
||||||
|
self._run_stack_build_command_from_build_config(build_config)
|
||||||
return
|
return
|
||||||
|
|
||||||
with open(args.config, "r") as f:
|
with open(args.config, "r") as f:
|
||||||
|
|
|
@ -184,7 +184,13 @@ class DistributionSpec(BaseModel):
|
||||||
)
|
)
|
||||||
docker_image: Optional[str] = None
|
docker_image: Optional[str] = None
|
||||||
providers: Dict[str, Union[str, List[str]]] = Field(
|
providers: Dict[str, Union[str, List[str]]] = Field(
|
||||||
default_factory=dict,
|
default={
|
||||||
|
"inference": "meta-reference",
|
||||||
|
"memory": "meta-reference",
|
||||||
|
"safety": "meta-reference",
|
||||||
|
"agents": "meta-reference",
|
||||||
|
"telemetry": "meta-reference",
|
||||||
|
},
|
||||||
description="""
|
description="""
|
||||||
Provider Types for each of the APIs provided by this distribution. If you
|
Provider Types for each of the APIs provided by this distribution. If you
|
||||||
select multiple providers, you should provide an appropriate 'routing_map'
|
select multiple providers, you should provide an appropriate 'routing_map'
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue