From 799f24b2d1856b47456056879517f6ab0dc75e76 Mon Sep 17 00:00:00 2001 From: yoonlee666 Date: Tue, 16 Jun 2020 10:00:04 +0800 Subject: [PATCH] add warmup_steps param check --- mindspore/nn/optim/adam.py | 1 + 1 file changed, 1 insertion(+) diff --git a/mindspore/nn/optim/adam.py b/mindspore/nn/optim/adam.py index 9586a404f9..d08cb443b9 100755 --- a/mindspore/nn/optim/adam.py +++ b/mindspore/nn/optim/adam.py @@ -391,6 +391,7 @@ class AdamWeightDecayDynamicLR(Optimizer): raise RuntimeError(f"The {self.cls_name} optimizer cannot support group setting.") _check_param_value(beta1, beta2, eps, weight_decay, self.cls_name) _check_learning_rate_value(learning_rate, end_learning_rate, decay_steps, power, self.cls_name) + validator.check_integer('warmup_steps', warmup_steps, 0, Rel.GE, self.cls_name) # turn them to scalar when me support scalar/tensor mix operations self.global_step = Parameter(initializer(0, [1]), name="global_step") self.warmup_steps = Tensor(np.array([warmup_steps]).astype(np.float32))