mirror of
https://github.com/huggingface/transformers.git
synced 2025-10-20 17:13:56 +08:00
@ -298,7 +298,7 @@ class TrainingArguments:
|
||||
`max_steps` is reached.
|
||||
lr_scheduler_type (`str` or [`SchedulerType`], *optional*, defaults to `"linear"`):
|
||||
The scheduler type to use. See the documentation of [`SchedulerType`] for all possible values.
|
||||
lr_scheduler_kwargs ('dict', *optional*, defaults to {}):
|
||||
lr_scheduler_kwargs (`dict` or `str`, *optional*, defaults to `None`):
|
||||
The extra arguments for the lr_scheduler. See the documentation of each scheduler for possible values.
|
||||
warmup_ratio (`float`, *optional*, defaults to 0.0):
|
||||
Ratio of total training steps used for a linear warmup from 0 to `learning_rate`.
|
||||
@ -880,8 +880,8 @@ class TrainingArguments:
|
||||
default="linear",
|
||||
metadata={"help": "The scheduler type to use."},
|
||||
)
|
||||
lr_scheduler_kwargs: Union[dict[str, Any], str] = field(
|
||||
default_factory=dict,
|
||||
lr_scheduler_kwargs: Optional[Union[dict, str]] = field(
|
||||
default=None,
|
||||
metadata={
|
||||
"help": (
|
||||
"Extra parameters for the lr_scheduler such as {'num_cycles': 1} for the cosine with hard restarts."
|
||||
|
Reference in New Issue
Block a user