mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-14 17:16:09 +00:00
fix: llama stack run starter in conda (#2679)
# What does this PR do? `llama stack run starter` in conda environment fails with ' --config is required for venv and conda environments' because it is passed as --template and start_stack.sh doesn't process template. ## Test Plan `llama stack run starter`
This commit is contained in:
parent
7915551eee
commit
780b4c6eea
2 changed files with 40 additions and 34 deletions
|
@ -83,46 +83,57 @@ class StackRun(Subcommand):
|
||||||
return ImageType.CONDA.value, args.image_name
|
return ImageType.CONDA.value, args.image_name
|
||||||
return args.image_type, args.image_name
|
return args.image_type, args.image_name
|
||||||
|
|
||||||
|
def _resolve_config_and_template(self, args: argparse.Namespace) -> tuple[Path | None, str | None]:
|
||||||
|
"""Resolve config file path and template name from args.config"""
|
||||||
|
from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR
|
||||||
|
|
||||||
|
if not args.config:
|
||||||
|
return None, None
|
||||||
|
|
||||||
|
config_file = Path(args.config)
|
||||||
|
has_yaml_suffix = args.config.endswith(".yaml")
|
||||||
|
template_name = None
|
||||||
|
|
||||||
|
if not config_file.exists() and not has_yaml_suffix:
|
||||||
|
# check if this is a template
|
||||||
|
config_file = Path(REPO_ROOT) / "llama_stack" / "templates" / args.config / "run.yaml"
|
||||||
|
if config_file.exists():
|
||||||
|
template_name = args.config
|
||||||
|
|
||||||
|
if not config_file.exists() and not has_yaml_suffix:
|
||||||
|
# check if it's a build config saved to ~/.llama dir
|
||||||
|
config_file = Path(DISTRIBS_BASE_DIR / f"llamastack-{args.config}" / f"{args.config}-run.yaml")
|
||||||
|
|
||||||
|
if not config_file.exists():
|
||||||
|
self.parser.error(
|
||||||
|
f"File {str(config_file)} does not exist.\n\nPlease run `llama stack build` to generate (and optionally edit) a run.yaml file"
|
||||||
|
)
|
||||||
|
|
||||||
|
if not config_file.is_file():
|
||||||
|
self.parser.error(
|
||||||
|
f"Config file must be a valid file path, '{config_file}' is not a file: type={type(config_file)}"
|
||||||
|
)
|
||||||
|
|
||||||
|
return config_file, template_name
|
||||||
|
|
||||||
def _run_stack_run_cmd(self, args: argparse.Namespace) -> None:
|
def _run_stack_run_cmd(self, args: argparse.Namespace) -> None:
|
||||||
import yaml
|
import yaml
|
||||||
|
|
||||||
from llama_stack.distribution.configure import parse_and_maybe_upgrade_config
|
from llama_stack.distribution.configure import parse_and_maybe_upgrade_config
|
||||||
from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR
|
|
||||||
from llama_stack.distribution.utils.exec import formulate_run_args, run_command
|
from llama_stack.distribution.utils.exec import formulate_run_args, run_command
|
||||||
|
|
||||||
if args.enable_ui:
|
if args.enable_ui:
|
||||||
self._start_ui_development_server(args.port)
|
self._start_ui_development_server(args.port)
|
||||||
image_type, image_name = self._get_image_type_and_name(args)
|
image_type, image_name = self._get_image_type_and_name(args)
|
||||||
|
|
||||||
|
# Resolve config file and template name first
|
||||||
|
config_file, template_name = self._resolve_config_and_template(args)
|
||||||
|
|
||||||
# Check if config is required based on image type
|
# Check if config is required based on image type
|
||||||
if (image_type in [ImageType.CONDA.value, ImageType.VENV.value]) and not args.config:
|
if (image_type in [ImageType.CONDA.value, ImageType.VENV.value]) and not config_file:
|
||||||
self.parser.error("Config file is required for venv and conda environments")
|
self.parser.error("Config file is required for venv and conda environments")
|
||||||
|
|
||||||
if args.config:
|
if config_file:
|
||||||
config_file = Path(args.config)
|
|
||||||
has_yaml_suffix = args.config.endswith(".yaml")
|
|
||||||
template_name = None
|
|
||||||
|
|
||||||
if not config_file.exists() and not has_yaml_suffix:
|
|
||||||
# check if this is a template
|
|
||||||
config_file = Path(REPO_ROOT) / "llama_stack" / "templates" / args.config / "run.yaml"
|
|
||||||
if config_file.exists():
|
|
||||||
template_name = args.config
|
|
||||||
|
|
||||||
if not config_file.exists() and not has_yaml_suffix:
|
|
||||||
# check if it's a build config saved to ~/.llama dir
|
|
||||||
config_file = Path(DISTRIBS_BASE_DIR / f"llamastack-{args.config}" / f"{args.config}-run.yaml")
|
|
||||||
|
|
||||||
if not config_file.exists():
|
|
||||||
self.parser.error(
|
|
||||||
f"File {str(config_file)} does not exist.\n\nPlease run `llama stack build` to generate (and optionally edit) a run.yaml file"
|
|
||||||
)
|
|
||||||
|
|
||||||
if not config_file.is_file():
|
|
||||||
self.parser.error(
|
|
||||||
f"Config file must be a valid file path, '{config_file}' is not a file: type={type(config_file)}"
|
|
||||||
)
|
|
||||||
|
|
||||||
logger.info(f"Using run configuration: {config_file}")
|
logger.info(f"Using run configuration: {config_file}")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
@ -138,8 +149,6 @@ class StackRun(Subcommand):
|
||||||
self.parser.error(f"failed to parse config file '{config_file}':\n {e}")
|
self.parser.error(f"failed to parse config file '{config_file}':\n {e}")
|
||||||
else:
|
else:
|
||||||
config = None
|
config = None
|
||||||
config_file = None
|
|
||||||
template_name = None
|
|
||||||
|
|
||||||
# If neither image type nor image name is provided, assume the server should be run directly
|
# If neither image type nor image name is provided, assume the server should be run directly
|
||||||
# using the current environment packages.
|
# using the current environment packages.
|
||||||
|
@ -172,10 +181,7 @@ class StackRun(Subcommand):
|
||||||
run_args.extend([str(args.port)])
|
run_args.extend([str(args.port)])
|
||||||
|
|
||||||
if config_file:
|
if config_file:
|
||||||
if template_name:
|
run_args.extend(["--config", str(config_file)])
|
||||||
run_args.extend(["--template", str(template_name)])
|
|
||||||
else:
|
|
||||||
run_args.extend(["--config", str(config_file)])
|
|
||||||
|
|
||||||
if args.env:
|
if args.env:
|
||||||
for env_var in args.env:
|
for env_var in args.env:
|
||||||
|
|
|
@ -197,7 +197,7 @@ def llama_stack_client(request, provider_data):
|
||||||
server_process = start_llama_stack_server(config_name)
|
server_process = start_llama_stack_server(config_name)
|
||||||
|
|
||||||
# Wait for server to be ready
|
# Wait for server to be ready
|
||||||
if not wait_for_server_ready(base_url, timeout=30, process=server_process):
|
if not wait_for_server_ready(base_url, timeout=120, process=server_process):
|
||||||
print("Server failed to start within timeout")
|
print("Server failed to start within timeout")
|
||||||
server_process.terminate()
|
server_process.terminate()
|
||||||
raise RuntimeError(
|
raise RuntimeError(
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue