Cross entropy - Wikipedia
https://en.wikipedia.org/wiki/Cross_entropyThe cross-entropy of the distribution relative to a distribution over a given set is defined as follows: ,where is the expected value operator with respect to the distribution . The definition may be formulated using the Kullback–Leibler divergence , divergence of from (also known as the relative entropy of with respect to ).
[2103.12532] Balanced Softmax Cross-Entropy for ...
https://arxiv.org/abs/2103.1253223.03.2021 · Balanced Softmax Cross-Entropy for Incremental Learning. Deep neural networks are prone to catastrophic forgetting when incrementally trained on new classes or new tasks as adaptation to the new data leads to a drastic decrease of the performance on the old classes and tasks. By using a small memory for rehearsal and knowledge distillation ...