Groups
Category
L2 regularization (also called ridge or weight decay) adds a penalty proportional to the sum of squared weights to discourage large parameters.
The biasโvariance tradeoff explains how prediction error splits into bias squared, variance, and irreducible noise for squared loss.