mindspore.mint.nn.functional.hardshrink

View Source On Gitee
mindspore.mint.nn.functional.hardshrink(input, lambd=0.5)[source]

Hard Shrink activation function. Calculates the output according to the input elements.

The formula is defined as follows:

\[\begin{split}\text{HardShrink}(x) = \begin{cases} x, & \text{ if } x > \lambda \\ x, & \text{ if } x < -\lambda \\ 0, & \text{ otherwise } \end{cases}\end{split}\]

HShrink Activation Function Graph:

../../_images/HShrink.png
Parameters
  • input (Tensor) –

    The input of Hard Shrink. Supported dtypes:

    • Ascend: float16, float32, bfloat16.

  • lambd (number, optional) – The threshold \(\lambda\) defined by the Hard Shrink formula. Default: 0.5 .

Returns

Tensor, has the same data type and shape as the input input.

Raises
  • TypeError – If lambd is not a float, int or bool.

  • TypeError – If input is not a tensor.

  • TypeError – If dtype of input is not float16, float32 or bfloat16.

Supported Platforms:

Ascend

Examples

>>> import mindspore
>>> import numpy as np
>>> from mindspore import Tensor, mint
>>> input = Tensor(np.array([[0.5, 1, 2.0], [0.0533, 0.0776, -2.1233]]), mindspore.float32)
>>> output = mint.nn.functional.hardshrink(input)
>>> print(output)
[[ 0.      1.      2.    ]
 [ 0.      0.     -2.1233]]