transformers-cli login => huggingface-cli login (#18490)

* zero chance anyone's using that constant no?

* `transformers-cli login` => `huggingface-cli login`

* `transformers-cli repo create` => `huggingface-cli repo create`

* `make style`
This commit is contained in:
Julien Chaumond
2022-08-06 09:42:55 +02:00
committed by GitHub
parent 8d1f9039d0
commit 9129fd0377
64 changed files with 72 additions and 75 deletions

View File

@ -186,7 +186,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -172,7 +172,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -173,7 +173,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -172,7 +172,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -172,7 +172,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -159,7 +159,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -186,7 +186,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -105,7 +105,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -153,7 +153,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -162,7 +162,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -156,7 +156,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -90,7 +90,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -145,7 +145,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -137,7 +137,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -157,7 +157,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -114,7 +114,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -111,7 +111,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -99,7 +99,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -83,7 +83,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -83,7 +83,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -82,7 +82,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -83,7 +83,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -246,7 +246,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -233,7 +233,7 @@ class DataTrainingArguments:
metadata={
"help": (
"If :obj:`True`, will use the token generated when running"
":obj:`transformers-cli login` as HTTP bearer authorization for remote files."
":obj:`huggingface-cli login` as HTTP bearer authorization for remote files."
)
},
)

View File

@ -89,7 +89,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -103,7 +103,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -192,7 +192,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -156,7 +156,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -83,7 +83,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -93,7 +93,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -81,7 +81,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -101,7 +101,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -84,7 +84,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -231,7 +231,7 @@ class DataTrainingArguments:
metadata={
"help": (
"If :obj:`True`, will use the token generated when running"
":obj:`transformers-cli login` as HTTP bearer authorization for remote files."
":obj:`huggingface-cli login` as HTTP bearer authorization for remote files."
)
},
)

View File

@ -234,7 +234,7 @@ class DataTrainingArguments:
metadata={
"help": (
"If :obj:`True`, will use the token generated when running"
":obj:`transformers-cli login` as HTTP bearer authorization for remote files."
":obj:`huggingface-cli login` as HTTP bearer authorization for remote files."
)
},
)

View File

@ -175,7 +175,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -104,7 +104,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -102,7 +102,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -287,7 +287,7 @@ class DataTrainingArguments:
metadata={
"help": (
"If :obj:`True`, will use the token generated when running"
":obj:`transformers-cli login` as HTTP bearer authorization for remote files."
":obj:`huggingface-cli login` as HTTP bearer authorization for remote files."
)
},
)

View File

@ -114,7 +114,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -115,7 +115,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -157,7 +157,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -80,7 +80,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -101,7 +101,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -183,7 +183,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -173,7 +173,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -83,7 +83,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -95,7 +95,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},

View File

@ -57,7 +57,7 @@ To upload all converted models,
2. Login to `transformers-cli`
```bash
transformers-cli login
huggingface-cli login
```
3. Run the `upload_models` script

View File

@ -2,7 +2,7 @@
for FILE in converted/*; do
model_name=`basename $FILE`
transformers-cli repo create $model_name -y
huggingface-cli repo create $model_name -y
git clone https://huggingface.co/Helsinki-NLP/$model_name
mv $FILE/* $model_name/
cd $model_name

View File

@ -22,9 +22,6 @@ from requests.exceptions import HTTPError
from . import BaseTransformersCLICommand
UPLOAD_MAX_FILES = 15
class UserCommands(BaseTransformersCLICommand):
@staticmethod
def register_subcommand(parser: ArgumentParser):
@ -105,7 +102,7 @@ class LoginCommand(BaseUserCommand):
def run(self):
print(
ANSI.red(
"ERROR! `transformers-cli login` uses an outdated login mechanism "
"ERROR! `huggingface-cli login` uses an outdated login mechanism "
"that is not compatible with the Hugging Face Hub backend anymore. "
"Please use `huggingface-cli login instead."
)

View File

@ -463,7 +463,7 @@ class PretrainedConfig(PushToHubMixin):
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -195,7 +195,7 @@ def get_cached_module_file(
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any
@ -345,7 +345,7 @@ def get_class_from_dynamic_module(
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or `bool`, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -251,7 +251,7 @@ class FeatureExtractionMixin(PushToHubMixin):
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -2096,7 +2096,7 @@ class TFPreTrainedModel(tf.keras.Model, TFModelUtilsMixin, TFGenerationMixin, Pu
Whether or not to only look at local files (e.g., not try doanloading the model).
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any
@ -2472,8 +2472,8 @@ class TFPreTrainedModel(tf.keras.Model, TFModelUtilsMixin, TFGenerationMixin, Pu
Whether or not the repository created should be private (requires a paying subscription).
use_auth_token (`bool` or `str`, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`). Will default to `True` if
`repo_url` is not specified.
when running `huggingface-cli login` (stored in `~/.huggingface`). Will default to `True` if `repo_url`
is not specified.
max_shard_size (`int` or `str`, *optional*, defaults to `"10GB"`):
Only applicable for models. The maximum size for a checkpoint before being sharded. Checkpoints shard
will then be each of size lower than this size. If expressed as a string, needs to be digits followed

View File

@ -1659,7 +1659,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin, GenerationMixin, PushToHubMix
Whether or not to only look at local files (i.e., do not try to download the model).
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -142,7 +142,7 @@ def get_feature_extractor_config(
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any
@ -247,7 +247,7 @@ class AutoFeatureExtractor:
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -135,7 +135,7 @@ class AutoProcessor:
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -357,7 +357,7 @@ def get_tokenizer_config(
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any

View File

@ -505,7 +505,7 @@ def pipeline(
Whether or not to use a Fast tokenizer if possible (a [`PreTrainedTokenizerFast`]).
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
device_map (`str` or `Dict[str, Union[int, str, torch.device]`, *optional*):
Sent directly as `model_kwargs` (just a simpler shortcut). When `accelerate` library is present, set
`device_map="auto"` to compute the most optimized `device_map` automatically. [More

View File

@ -1596,7 +1596,7 @@ class PreTrainedTokenizerBase(SpecialTokensMixin, PushToHubMixin):
'http://hostname': 'foo.bar:4012'}`. The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
local_files_only (`bool`, *optional*, defaults to `False`):
Whether or not to only rely on local files and not to attempt to download any files.
revision (`str`, *optional*, defaults to `"main"`):

View File

@ -716,7 +716,7 @@ def cached_file(
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any
@ -870,7 +870,7 @@ def get_file_from_repo(
'http://hostname': 'foo.bar:4012'}.` The proxies are used on each request.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
revision (`str`, *optional*, defaults to `"main"`):
The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a
git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any
@ -983,7 +983,7 @@ def get_list_of_files(
identifier allowed by git.
use_auth_token (`str` or *bool*, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`).
when running `huggingface-cli login` (stored in `~/.huggingface`).
local_files_only (`bool`, *optional*, defaults to `False`):
Whether or not to only rely on local files and not to attempt to download any files.
@ -1161,8 +1161,8 @@ class PushToHubMixin:
Whether or not the repository created should be private (requires a paying subscription).
use_auth_token (`bool` or `str`, *optional*):
The token to use as HTTP bearer authorization for remote files. If `True`, will use the token generated
when running `transformers-cli login` (stored in `~/.huggingface`). Will default to `True` if
`repo_url` is not specified.
when running `huggingface-cli login` (stored in `~/.huggingface`). Will default to `True` if `repo_url`
is not specified.
max_shard_size (`int` or `str`, *optional*, defaults to `"10GB"`):
Only applicable for models. The maximum size for a checkpoint before being sharded. Checkpoints shard
will then be each of size lower than this size. If expressed as a string, needs to be digits followed

View File

@ -118,7 +118,7 @@ class ModelArguments:
use_auth_token: bool = field(
default=False,
metadata={
"help": "Will use the token generated when running `transformers-cli login` (necessary to use this script "
"help": "Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
},
)

View File

@ -181,7 +181,7 @@ class ModelArguments:
default=False,
metadata={
"help": (
"Will use the token generated when running `transformers-cli login` (necessary to use this script "
"Will use the token generated when running `huggingface-cli login` (necessary to use this script "
"with private models)."
)
},