Categorical Cross-Entropy Loss is a popular cost function used in multi-class classification problems. It calculates the error between the predicted probability distribution and the actual probability distribution of the classes.
The formula for categorical cross-entropy loss is:
This can be simplified further when one hot encoding is used.
Where
In some algorithms it's possible for
This means that the loss function can then be re-written formally as:
Derivative
Again a small