Browse Source

!6115 API comment enhancement & ckpt bug fix

Merge pull request !6115 from Simson/push-to-opensource
tags/v1.0.0
mindspore-ci-bot Gitee 5 years ago
parent
commit
bca498eb53
4 changed files with 19 additions and 19 deletions
  1. +11
    -11
      mindspore/ops/operations/nn_ops.py
  2. +1
    -1
      mindspore/ops/operations/other_ops.py
  3. +2
    -2
      mindspore/ops/operations/random_ops.py
  4. +5
    -5
      mindspore/train/serialization.py

+ 11
- 11
mindspore/ops/operations/nn_ops.py View File

@@ -149,7 +149,7 @@ class LogSoftmax(PrimitiveWithInfer):
Log Softmax activation function.

Applies the Log Softmax function to the input tensor on the specified axis.
Suppose a slice in the given aixs :math:`x` then for each element :math:`x_i`
Suppose a slice in the given aixs, :math:`x` for each element :math:`x_i`,
the Log Softmax function is shown as follows:

.. math::
@@ -158,7 +158,7 @@ class LogSoftmax(PrimitiveWithInfer):
where :math:`N` is the length of the Tensor.

Args:
axis (int): The axis to do the Log softmax operation. Default: -1.
axis (int): The axis to perform the Log softmax operation. Default: -1.

Inputs:
- **logits** (Tensor) - The input of Log Softmax, with float16 or float32 data type.
@@ -2253,7 +2253,7 @@ class L2Normalize(PrimitiveWithInfer):
r"""
L2 normalization Operator.

This operator will normalizes the input using the given axis. The function is shown as follows:
This operator will normalize the input using the given axis. The function is shown as follows:

.. math::
\text{output} = \frac{x}{\sqrt{\text{max}(\text{sum} (\text{input_x}^2), \epsilon)}},
@@ -2261,7 +2261,7 @@ class L2Normalize(PrimitiveWithInfer):
where :math:`\epsilon` is epsilon.

Args:
axis (int): The begin axis for the input to apply L2 normalize. Default: 0.
axis (int): The starting axis for the input to apply the L2 normalization. Default: 0.
epsilon (float): A small value added for numerical stability. Default: 1e-4.

Inputs:
@@ -2657,7 +2657,7 @@ class LSTM(PrimitiveWithInfer):
"""
Performs the long short term memory(LSTM) on the input.

Detailed information, please refer to `nn.LSTM`.
For detailed information, please refer to `nn.LSTM`.
"""

@prim_attr_register
@@ -4803,7 +4803,7 @@ class ApplyProximalGradientDescent(PrimitiveWithInfer):

class LARSUpdate(PrimitiveWithInfer):
"""
Conduct lars (layer-wise adaptive rate scaling) update on the square sum of gradient.
Conduct lars (layer-wise adaptive rate scaling) update on the sum of squares of gradient.

Args:
epsilon (float): Term added to the denominator to improve numerical stability. Default: 1e-05.
@@ -4813,8 +4813,8 @@ class LARSUpdate(PrimitiveWithInfer):
Inputs:
- **weight** (Tensor) - The weight to be updated.
- **gradient** (Tensor) - The gradient of weight, which has the same shape and dtype with weight.
- **norm_weight** (Tensor) - A scalar tensor, representing the square sum of weight.
- **norm_gradient** (Tensor) - A scalar tensor, representing the square sum of gradient.
- **norm_weight** (Tensor) - A scalar tensor, representing the sum of squares of weight.
- **norm_gradient** (Tensor) - A scalar tensor, representing the sum of squares of gradient.
- **weight_decay** (Union[Number, Tensor]) - Weight decay. It should be a scalar tensor or number.
- **learning_rate** (Union[Number, Tensor]) - Learning rate. It should be a scalar tensor or number.

@@ -5576,10 +5576,10 @@ class InTopK(PrimitiveWithInfer):

class LRN(PrimitiveWithInfer):
r"""
Local Response Normalization
Local Response Normalization.

Args:
depth_radius (int): Half-width of the 1-D normalization window. Shape of 0-D.
depth_radius (int): Half-width of the 1-D normalization window with the shape of 0-D.
bias (float): An offset (usually positive to avoid dividing by 0).
alpha (float): A scale factor, usually positive.
beta (float): An exponent.
@@ -5589,7 +5589,7 @@ class LRN(PrimitiveWithInfer):
- **x** (Tensor) - A 4D Tensor with float16 or float32 data type.

Outputs:
Tensor, With shape and data type same as the input tensor.
Tensor, with the same shape and data type as the input tensor.

Examples:
>>> x = Tensor(np.random.rand(1, 10, 4, 4)), mindspore.float32)


+ 1
- 1
mindspore/ops/operations/other_ops.py View File

@@ -295,7 +295,7 @@ class MakeRefKey(Primitive):
tag (str): Parameter name to make the RefKey.

Inputs:
No input.
No inputs.

Outputs:
RefKeyType, made from the Parameter name.


+ 2
- 2
mindspore/ops/operations/random_ops.py View File

@@ -72,7 +72,7 @@ class Laplace(PrimitiveWithInfer):
\text{f}(x;μ,λ) = \frac{1}{2λ}\exp(-\frac{|x-μ|}{λ}),

Args:
seed (int): Seed data is used as entropy source for Random number engines generating pseudo-random numbers.
seed (int): Seed data is used as entropy source for Random number engines to generate pseudo-random numbers.
Default: 0.

Inputs:
@@ -83,7 +83,7 @@ class Laplace(PrimitiveWithInfer):
variance of Laplace distribution is equal to twice the square of lambda_param. With float32 data type.

Outputs:
Tensor, has the shape 'shape' input and dtype as float32.
Tensor, has the specified shape and its dtype is float32.

Examples:
>>> shape = (4, 16)


+ 5
- 5
mindspore/train/serialization.py View File

@@ -138,7 +138,7 @@ def _exec_save(ckpt_file_name, data_list):

except BaseException as e:
logger.error("Failed to save the checkpoint file %s.", ckpt_file_name)
raise RuntimeError(e.__str__())
raise e


def save_checkpoint(save_obj, ckpt_file_name, integrated_save=True, async_save=False):
@@ -146,10 +146,10 @@ def save_checkpoint(save_obj, ckpt_file_name, integrated_save=True, async_save=F
Saves checkpoint info to a specified file.

Args:
save_obj (nn.Cell or list): The train network for training or parameters list(each element is a dictionary,
like {"name":xx, "data":xx}.)
ckpt_file_name (str): Checkpoint file name.
integrated_save (bool): Whether to integrated save in automatic model parallel scene. Default: True
save_obj (nn.Cell or list): The cell object or parameters list(each element is a dictionary,
like {"name": param_name, "data": param_data}.)
ckpt_file_name (str): Checkpoint file name. If the file name already exists, it will be overwritten.
integrated_save (bool): Whether to integrated save in automatic model parallel scene.
async_save (bool): Whether asynchronous execution saves the checkpoint to a file. Default: False

Raises:


Loading…
Cancel
Save