mindspore.mint.nn.SELU
- class mindspore.mint.nn.SELU[source]
Activation function SELU (Scaled exponential Linear Unit).
Refer to
mindspore.mint.nn.functional.selu()
for more details.SELU Activation Function Graph:
- Supported Platforms:
Ascend
Examples
>>> import mindspore >>> from mindspore import Tensor, mint >>> import numpy as np >>> input = Tensor(np.array([[-1.0, 4.0, -8.0], [2.0, -5.0, 9.0]]), mindspore.float32) >>> selu = mint.nn.SELU() >>> output = selu(input) >>> print(output) [[-1.1113307 4.202804 -1.7575096] [ 2.101402 -1.7462534 9.456309 ]]