You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

utils.py 3.9 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105
  1. # Copyright 2020 Huawei Technologies Co., Ltd
  2. #
  3. # Licensed under the Apache License, Version 2.0 (the "License");
  4. # you may not use this file except in compliance with the License.
  5. # You may obtain a copy of the License at
  6. #
  7. # http://www.apache.org/licenses/LICENSE-2.0
  8. #
  9. # Unless required by applicable law or agreed to in writing, software
  10. # distributed under the License is distributed on an "AS IS" BASIS,
  11. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. # See the License for the specific language governing permissions and
  13. # limitations under the License.
  14. # ============================================================================
  15. """Resnet50 utils"""
  16. import time
  17. import numpy as np
  18. from mindspore.train.callback import Callback
  19. from mindspore import Tensor
  20. from mindspore import nn
  21. from mindspore.nn.loss.loss import LossBase
  22. from mindspore.ops import operations as P
  23. from mindspore.ops import functional as F
  24. from mindspore.common import dtype as mstype
  25. class Monitor(Callback):
  26. """
  27. Monitor loss and time.
  28. Args:
  29. lr_init (numpy array): train lr
  30. Returns:
  31. None
  32. Examples:
  33. >>> Monitor(100,lr_init=Tensor([0.05]*100).asnumpy())
  34. """
  35. def __init__(self, lr_init=None, step_threshold=10):
  36. super(Monitor, self).__init__()
  37. self.lr_init = lr_init
  38. self.lr_init_len = len(lr_init)
  39. self.step_threshold = step_threshold
  40. def epoch_begin(self, run_context):
  41. self.losses = []
  42. self.epoch_time = time.time()
  43. def epoch_end(self, run_context):
  44. cb_params = run_context.original_args()
  45. epoch_mseconds = (time.time() - self.epoch_time) * 1000
  46. per_step_mseconds = epoch_mseconds / cb_params.batch_num
  47. print("epoch time: {:5.3f}, per step time: {:5.3f}, avg loss: {:8.6f}".format(epoch_mseconds,
  48. per_step_mseconds,
  49. np.mean(self.losses)))
  50. self.epoch_mseconds = epoch_mseconds
  51. def step_begin(self, run_context):
  52. self.step_time = time.time()
  53. def step_end(self, run_context):
  54. cb_params = run_context.original_args()
  55. step_mseconds = (time.time() - self.step_time) * 1000
  56. step_loss = cb_params.net_outputs
  57. if isinstance(step_loss, (tuple, list)) and isinstance(step_loss[0], Tensor):
  58. step_loss = step_loss[0]
  59. if isinstance(step_loss, Tensor):
  60. step_loss = np.mean(step_loss.asnumpy())
  61. self.losses.append(step_loss)
  62. cur_step_in_epoch = (cb_params.cur_step_num - 1) % cb_params.batch_num
  63. print("epoch: [{:3d}/{:3d}], step:[{:5d}/{:5d}], loss:[{:8.6f}/{:8.6f}], time:[{:5.3f}], lr:[{:5.5f}]".format(
  64. cb_params.cur_epoch_num, cb_params.epoch_num, cur_step_in_epoch +
  65. 1, cb_params.batch_num, step_loss,
  66. np.mean(self.losses), step_mseconds, self.lr_init[cb_params.cur_step_num - 1]))
  67. if cb_params.cur_step_num == self.step_threshold:
  68. run_context.request_stop()
  69. class CrossEntropy(LossBase):
  70. """the redefined loss function with SoftmaxCrossEntropyWithLogits"""
  71. def __init__(self, smooth_factor=0, num_classes=1001):
  72. super(CrossEntropy, self).__init__()
  73. self.onehot = P.OneHot()
  74. self.on_value = Tensor(1.0 - smooth_factor, mstype.float32)
  75. self.off_value = Tensor(1.0 * smooth_factor /
  76. (num_classes - 1), mstype.float32)
  77. self.ce = nn.SoftmaxCrossEntropyWithLogits()
  78. self.mean = P.ReduceMean(False)
  79. def construct(self, logit, label):
  80. one_hot_label = self.onehot(label, F.shape(
  81. logit)[1], self.on_value, self.off_value)
  82. loss = self.ce(logit, one_hot_label)
  83. loss = self.mean(loss, 0)
  84. return loss