From 378a2077dde5d5a96da5ccd8a1f4b4359655b5fc Mon Sep 17 00:00:00 2001 From: Lucain Date: Thu, 25 Jul 2024 01:50:40 +0200 Subject: [PATCH] Update download command (#9) --- llama_toolchain/cli/download.py | 8 +++----- 1 file changed, 3 insertions(+), 5 deletions(-) diff --git a/llama_toolchain/cli/download.py b/llama_toolchain/cli/download.py index c126fbf37..ed41ac505 100644 --- a/llama_toolchain/cli/download.py +++ b/llama_toolchain/cli/download.py @@ -48,13 +48,13 @@ class Download(Subcommand): self.parser.add_argument( "repo_id", type=str, - help="Name of the repository on Hugging Face Hub eg. llhf/Meta-Llama-3.1-70B-Instruct", + help="Name of the repository on Hugging Face Hub eg. meta-llama/Meta-Llama-3.1-70B-Instruct", ) self.parser.add_argument( "--hf-token", type=str, required=False, - default=os.getenv("HF_TOKEN", None), + default=None, help="Hugging Face API token. Needed for gated models like Llama2. Will also try to read environment variable `HF_TOKEN` as default.", ) self.parser.add_argument( @@ -74,11 +74,9 @@ class Download(Subcommand): true_output_dir = snapshot_download( args.repo_id, local_dir=output_dir, - # "auto" will download to cache_dir and symlink files to local_dir - # avoiding unnecessary duplicate copies - local_dir_use_symlinks="auto", ignore_patterns=args.ignore_patterns, token=args.hf_token, + library_name="llama-toolchain", ) except GatedRepoError: self.parser.error(