mindspore.mint.nn.functional.binary_cross_entropy
- mindspore.mint.nn.functional.binary_cross_entropy(logits, labels, weight=None, reduction='mean')[source]
Computes the binary cross entropy(Measure the difference information between two probability distributions) between predictive value logits and target value labels.
Set logits as \(x\), labels as \(y\), output as \(\ell(x, y)\), the weight of nth batch of binary cross entropy is \(w_n\). Let,
\[L = \{l_1,\dots,l_N\}^\top, \quad l_n = - w_n \left[ y_n \cdot \log x_n + (1 - y_n) \cdot \log (1 - x_n) \right]\]In which, \(L\) indicates the loss of all batch_size, \(l\) indicates the loss of one batch_size, and \(n\) indicates one batch_size in the \(1-N\) range. Then,
\[\begin{split}\ell(x, y) = \begin{cases} L, & \text{if reduction} = \text{'none';}\\ \operatorname{mean}(L), & \text{if reduction} = \text{'mean';}\\ \operatorname{sum}(L), & \text{if reduction} = \text{'sum'.} \end{cases}\end{split}\]Warning
The value of logits must range from 0 to l.
- Parameters
logits (Tensor) – The predictive value whose data type must be float16 or float32.
labels (Tensor) – The target value which has the same shape and data type as logits. And the data type is float16 or float32.
weight (Tensor, optional) – A rescaling weight applied to the loss of each batch element. Its shape must be able to broadcast to that of logits and labels. And it must have the same shape and data type as logits. Default:
None
. If set toNone
, the loss function will not consider any sample weights, and each sample will be treated as having equal importance when calculating the loss.reduction (str, optional) –
Apply specific reduction method to the output:
'none'
,'mean'
,'sum'
. Default:'mean'
.'none'
: no reduction will be applied.'mean'
: compute and return the weighted mean of elements in the output.'sum'
: the output elements will be summed.
- Returns
Tensor or Scalar. Returns Tensor that has the same dtype and shape as logits if reduction is 'none'. Otherwise, returns a scalar Tensor.
- Raises
TypeError – If logits, labels or weight is not a Tensor.
TypeError – If dtype of logits, labels or weight (if given) is neither float16 nor float32.
ValueError – If reduction is not one of
'none'
,'mean'
or'sum'
.ValueError – If shape of labels is not the same as logits or weight (if given).
- Supported Platforms:
Ascend
Examples
>>> import mindspore >>> import numpy as np >>> from mindspore import Tensor, mint >>> logits = Tensor(np.array([0.2, 0.7, 0.1]), mindspore.float32) >>> labels = Tensor(np.array([0., 1., 0.]), mindspore.float32) >>> weight = Tensor(np.array([1, 2, 2]), mindspore.float32) >>> output = mint.nn.functional.binary_cross_entropy(logits, labels, weight) >>> print(output) 0.38240486