Groups
Cross-entropy loss measures how well predicted probabilities match the true labels by penalizing confident wrong predictions heavily.
The Lottery Ticket Hypothesis (LTH) says that inside a large dense neural network there exist small sparse subnetworks that, when trained in isolation from their original initialization, can reach comparable accuracy to the full model.