Groups
Category
Level
Early stopping halts training when the validation loss stops improving, preventing overfitting and saving compute.
Dropout randomly turns off (zeros) some neurons during training to prevent the network from memorizing the training data.
L2 regularization (also called ridge or weight decay) adds a penalty proportional to the sum of squared weights to discourage large parameters.