{\mu}P$^2$: Effective Sharpness Aware Minimization Requires Layerwise Perturbation Scaling
Sharpness Aware Minimization (SAM) enhances performance across various neural architectures and datasets. As models are continually scaled up to improve performance, a rigorous understanding of SAM's scaling behaviour is paramount. To this end, we study the infinite-width limit of neural networ...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Sharpness Aware Minimization (SAM) enhances performance across various neural
architectures and datasets. As models are continually scaled up to improve
performance, a rigorous understanding of SAM's scaling behaviour is paramount.
To this end, we study the infinite-width limit of neural networks trained with
SAM, using the Tensor Programs framework. Our findings reveal that the dynamics
of standard SAM effectively reduce to applying SAM solely in the last layer in
wide neural networks, even with optimal hyperparameters. In contrast, we
identify a stable parameterization with layerwise perturbation scaling, which
we call $\textit{Maximal Update and Perturbation Parameterization}$
($\mu$P$^2$), that ensures all layers are both feature learning and effectively
perturbed in the limit. Through experiments with MLPs, ResNets and Vision
Transformers, we empirically demonstrate that $\mu$P$^2$ achieves
hyperparameter transfer of the joint optimum of learning rate and perturbation
radius across model scales. Moreover, we provide an intuitive condition to
derive $\mu$P$^2$ for other perturbation rules like Adaptive SAM and SAM-ON,
also ensuring balanced perturbation effects across all layers. |
---|---|
DOI: | 10.48550/arxiv.2411.00075 |