Cross Entropy Loss
The cross entropy between two probability distributions over the same underlying set of events measures the average number of bits needed to identify an event drawn from the set.
Cross entropy can be used to calculate loss.
Regularization
Regularization is the process of introducing additional information to prevent overfitting and reduce loss, including:
L1 - Lasso Regression; variable selection and regularization
L2 - Ridge Regression; useful to mitigate multicollinearity
An Example
An example is: