mindspore.nn.LSTMCell
- class mindspore.nn.LSTMCell(input_size: int, hidden_size: int, has_bias: bool = True, dtype=mstype.float32)[source]
A LSTM (Long Short-Term Memory) cell.
Here
is the sigmoid function, and is the Hadamard product. are learnable weights between the output and the input in the formula. For instance, are the weight and bias used to transform from input to . Details can be found in paper LONG SHORT-TERM MEMORY and Long Short-Term Memory Recurrent Neural Network Architectures for Large Scale Acoustic Modeling.The encapsulated LSTMCell can be simplified to the following formula:
- Parameters
input_size (int) – Number of features of input.
hidden_size (int) – Number of features of hidden layer.
has_bias (bool) – Whether the cell has bias b_{ih} and b_{hh}. Default:
True
.dtype (
mindspore.dtype
) – Dtype of Parameters. Default:mstype.float32
.
- Inputs:
x (Tensor) - Tensor of shape
.hx (tuple) - A tuple of two Tensors (h_0, c_0) both of data type mindspore.float32 and shape
.
- Outputs:
hx' (Tensor) - A tuple of two Tensors (h', c') both of data shape
.
- Supported Platforms:
Ascend
GPU
CPU
Examples
>>> import mindspore as ms >>> import numpy as np >>> net = ms.nn.LSTMCell(10, 16) >>> x = ms.Tensor(np.ones([5, 3, 10]).astype(np.float32)) >>> h = ms.Tensor(np.ones([3, 16]).astype(np.float32)) >>> c = ms.Tensor(np.ones([3, 16]).astype(np.float32)) >>> output = [] >>> for i in range(5): ... hx = net(x[i], (h, c)) ... output.append(hx) >>> print(output[0][0].shape) (3, 16)