Groups
Category
Stochastic Depth randomly drops whole residual layers during training while keeping the full network at inference time.
Dropout randomly turns off (zeros) some neurons during training to prevent the network from memorizing the training data.