From d82a9d94e32358babe4819948e99b78051ca13de Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Wed, 25 Sep 2024 10:56:13 -0700 Subject: [PATCH] Small fix to the prompt-format error message --- llama_stack/cli/model/prompt_format.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/llama_stack/cli/model/prompt_format.py b/llama_stack/cli/model/prompt_format.py index 7b1084ee4..e6fd8aac7 100644 --- a/llama_stack/cli/model/prompt_format.py +++ b/llama_stack/cli/model/prompt_format.py @@ -56,14 +56,14 @@ class ModelPromptFormat(Subcommand): try: model_id = CoreModelId(args.model_name) except ValueError: - raise argparse.ArgumentTypeError( + self.parser.error( f"{args.model_name} is not a valid Model. Choose one from --\n{model_str}" - ) from None + ) if model_id not in supported_model_ids: - raise argparse.ArgumentTypeError( + self.parser.error( f"{model_id} is not a valid Model. Choose one from --\n {model_str}" - ) from None + ) llama_3_1_file = pkg_resources.resource_filename( "llama_models", "llama3_1/prompt_format.md"