mindspore.mint.nn.functional.mse_loss
- mindspore.mint.nn.functional.mse_loss(input, target, reduction='mean')[source]
Calculates the mean squared error between the predicted value and the label value.
For detailed information, please refer to
mindspore.nn.MSELoss
.- Parameters
input (Tensor) – Tensor of any dimension. The data type needs to be consistent with the target. It should also be broadcastable with the target.
target (Tensor) – The input label. Tensor of any dimension. The data type needs to be consistent with the input. It should also be broadcastable with the input.
reduction (str, optional) –
Apply specific reduction method to the output:
'mean'
,'none'
,'sum'
. Default:'mean'
.'none'
: no reduction will be applied.'mean'
: compute and return the mean of elements in the output.'sum'
: the output elements will be summed.
- Returns
Tensor. If reduction is
'mean'
or'sum'
, the shape of output is Tensor Scalar.If reduction is
'none'
, the shape of output is the broadcasted shape of input and target .
- Raises
ValueError – If reduction is not one of
'mean'
,'sum'
or'none'
.ValueError – If input and target are not broadcastable.
TypeError – If input and target are in different data type.
- Supported Platforms:
Ascend
Examples
>>> import mindspore >>> import numpy as np >>> from mindspore import Tensor, mint >>> logits = Tensor(np.array([1, 2, 3]), mindspore.float32) >>> labels = Tensor(np.array([[1, 1, 1], [1, 2, 2]]), mindspore.float32) >>> output = mint.nn.functional.mse_loss(logits, labels, reduction='none') >>> print(output) [[0. 1. 4.] [0. 0. 1.]]