diff --git a/mindspore/nn/layer/quant.py b/mindspore/nn/layer/quant.py index 1c55d86df6..15cf5b58c0 100644 --- a/mindspore/nn/layer/quant.py +++ b/mindspore/nn/layer/quant.py @@ -27,6 +27,7 @@ from mindspore.nn.cell import Cell from mindspore.nn.layer.activation import get_activation import mindspore.context as context + __all__ = [ 'FakeQuantWithMinMax', 'DepthwiseConv2dBatchNormQuant', @@ -130,7 +131,6 @@ class FakeQuantWithMinMaxD(Cell): >>> input_x = Tensor(np.array([[1, 2, 1], [-2, 0, -1]]), mindspore.float32) >>> result = fake_quant(input_x) """ - def __init__(self, min_init=-6, max_init=6, @@ -485,9 +485,9 @@ class DepthwiseConv2dBatchNormQuant(Cell): s = 'in_channels={}, out_channels={}, kernel_size={}, stride={}, ' \ 'pad_mode={}, padding={}, dilation={}, group={}, ' \ 'fake={}, freeze_bn={}, momentum={}, quant_delay={}'.format( - self.in_channels, self.out_channels, self.kernel_size, self.stride, - self.pad_mode, self.padding, self.dilation, self.group, - self.fake, self.freeze_bn, self.momentum, self.quant_delay) + self.in_channels, self.out_channels, self.kernel_size, self.stride, + self.pad_mode, self.padding, self.dilation, self.group, + self.fake, self.freeze_bn, self.momentum, self.quant_delay) return s def construct(self, x): @@ -662,9 +662,9 @@ class Conv2dBatchNormQuant(Cell): s = 'in_channels={}, out_channels={}, kernel_size={}, stride={}, ' \ 'pad_mode={}, padding={}, dilation={}, group={}, ' \ 'fake={}, freeze_bn={}, momentum={}, quant_delay={}'.format( - self.in_channels, self.out_channels, self.kernel_size, self.stride, - self.pad_mode, self.padding, self.dilation, self.group, - self.fake, self.freeze_bn, self.momentum, self.quant_delay) + self.in_channels, self.out_channels, self.kernel_size, self.stride, + self.pad_mode, self.padding, self.dilation, self.group, + self.fake, self.freeze_bn, self.momentum, self.quant_delay) return s def construct(self, x): @@ -807,9 +807,9 @@ class Conv2dQuant(Cell): s = 'in_channels={}, out_channels={}, kernel_size={}, stride={}, ' \ 'pad_mode={}, padding={}, dilation={}, group={}, ' \ 'has_bias={}, quant_delay={}'.format( - self.in_channels, self.out_channels, self.kernel_size, self.stride, - self.pad_mode, self.padding, self.dilation, self.group, - self.has_bias, self.quant_delay) + self.in_channels, self.out_channels, self.kernel_size, self.stride, + self.pad_mode, self.padding, self.dilation, self.group, + self.has_bias, self.quant_delay) return s diff --git a/mindspore/ops/_op_impl/_custom_op/__init__.py b/mindspore/ops/_op_impl/_custom_op/__init__.py index e2d6238be8..d111761cd8 100644 --- a/mindspore/ops/_op_impl/_custom_op/__init__.py +++ b/mindspore/ops/_op_impl/_custom_op/__init__.py @@ -23,4 +23,4 @@ from .correction_mul import _correction_mul_tbe from .correction_mul_grad import _correction_mul_grad_tbe from .fake_quant_with_min_max import _fake_quant_tbe from .fake_quant_with_min_max_grad import _fake_quant_grad_tbe -from .fake_quant_with_min_max_update import _fake_quant_update5d_tbe \ No newline at end of file +from .fake_quant_with_min_max_update import _fake_quant_update5d_tbe