Groups
Category
Sharpness-Aware Minimization (SAM) trains models to perform well even when their weights are slightly perturbed, seeking flatter minima that generalize better.
Multi-task loss balancing aims to automatically set each task’s weight so that no single loss dominates training.