mindspore.ops.GLU
- class mindspore.ops.GLU(axis=- 1)[source]
Computes GLU (Gated Linear Unit activation function) of input tensors.
Warning
This is an experimental API that is subject to change or deletion.
Refer to
mindspore.ops.glu()
for more details.- Parameters
axis (int, optional) – Axis on which to split the input. The value of axis must be an int within range [-rank(x), rank(x)). Default: -1, specifying the last dimension.
- Inputs:
x (Tensor) - Input tensor. x.shape[axis] must be even.
- Outputs:
Tensor, has the same data type with x.
- Supported Platforms:
Ascend
CPU
Examples
>>> from mindspore import ops, Tensor >>> from mindspore import dtype as mstype >>> import numpy as np >>> axis = 0 >>> x = Tensor(np.array([0.3220, 0.9545, 0.7879, 0.0975, 0.3698, ... 0.5135, 0.5740, 0.3435, 0.1895, 0.8764, ... 0.4980, 0.9673, 0.9879, 0.6988, 0.9022, ... 0.9304, 0.1558, 0.0153, 0.1559, 0.9852]).reshape([2, 2, 5]), mstype.float32) >>> glu = ops.GLU(axis=axis) >>> y = glu(x) >>> print(y) [[[0.20028052 0.6916126 0.57412136 0.06512236 0.26307625] [0.3682598 0.3093122 0.17306386 0.10212085 0.63814086]]]