Groups
Category
Stochastic Depth randomly drops whole residual layers during training while keeping the full network at inference time.
Batch Normalization rescales and recenters activations using mini-batch statistics to stabilize and speed up neural network training.