diff --git a/mindspore/nn/optim/lamb.py b/mindspore/nn/optim/lamb.py index 99f4b17978..997c5304d5 100755 --- a/mindspore/nn/optim/lamb.py +++ b/mindspore/nn/optim/lamb.py @@ -241,7 +241,7 @@ class Lamb(Optimizer): Examples: >>> net = Net() >>> #1) All parameters use the same learning rate and weight decay - >>> optim = nn.Lamb(params=net.trainable_params(learning_rate=0.1)) + >>> optim = nn.Lamb(params=net.trainable_params(), learning_rate=0.1) >>> >>> #2) Use parameter groups and set different values >>> poly_decay_lr = learning_rate_schedule.PolynomialDecayLR()