mirror of
https://github.com/huggingface/transformers.git
synced 2025-10-20 17:13:56 +08:00
@ -88,7 +88,6 @@ from transformers.testing_utils import (
|
||||
require_ray,
|
||||
require_schedulefree,
|
||||
require_sentencepiece,
|
||||
require_sigopt,
|
||||
require_tensorboard,
|
||||
require_tokenizers,
|
||||
require_torch,
|
||||
@ -5690,58 +5689,6 @@ class TrainerHyperParameterRayIntegrationTest(unittest.TestCase):
|
||||
self.ray_hyperparameter_search()
|
||||
|
||||
|
||||
@slow
|
||||
@require_torch
|
||||
@require_sigopt
|
||||
class TrainerHyperParameterSigOptIntegrationTest(unittest.TestCase):
|
||||
def setUp(self):
|
||||
args = TrainingArguments("..")
|
||||
self.n_epochs = args.num_train_epochs
|
||||
self.batch_size = args.train_batch_size
|
||||
|
||||
def test_hyperparameter_search(self):
|
||||
class MyTrialShortNamer(TrialShortNamer):
|
||||
DEFAULTS = {"a": 0, "b": 0}
|
||||
|
||||
def hp_space(trial):
|
||||
return [
|
||||
{"bounds": {"min": -4, "max": 4}, "name": "a", "type": "int"},
|
||||
{"bounds": {"min": -4, "max": 4}, "name": "b", "type": "int"},
|
||||
]
|
||||
|
||||
def model_init(trial):
|
||||
if trial is not None:
|
||||
a = trial.assignments["a"]
|
||||
b = trial.assignments["b"]
|
||||
else:
|
||||
a = 0
|
||||
b = 0
|
||||
config = RegressionModelConfig(a=a, b=b, double_output=False)
|
||||
|
||||
return RegressionPreTrainedModel(config).to(torch_device)
|
||||
|
||||
def hp_name(trial):
|
||||
return MyTrialShortNamer.shortname(trial.assignments)
|
||||
|
||||
with tempfile.TemporaryDirectory() as tmp_dir:
|
||||
trainer = get_regression_trainer(
|
||||
output_dir=tmp_dir,
|
||||
learning_rate=0.1,
|
||||
logging_steps=1,
|
||||
eval_strategy=IntervalStrategy.EPOCH,
|
||||
save_strategy=IntervalStrategy.EPOCH,
|
||||
num_train_epochs=4,
|
||||
disable_tqdm=True,
|
||||
load_best_model_at_end=True,
|
||||
logging_dir="runs",
|
||||
run_name="test",
|
||||
model_init=model_init,
|
||||
)
|
||||
trainer.hyperparameter_search(
|
||||
direction="minimize", hp_space=hp_space, hp_name=hp_name, backend="sigopt", n_trials=4
|
||||
)
|
||||
|
||||
|
||||
optim_test_params = []
|
||||
if is_torch_available():
|
||||
default_adam_kwargs = {
|
||||
|
Reference in New Issue
Block a user