Ada-boundary: accelerating DNN training via adaptive boundary batch selection
Neural networks converge faster with help from a smart batch selection strategy. In this regard, we propose Ada-Boundary , a novel and simple adaptive batch selection algorithm that constructs an effective mini-batch according to the learning progress of the model. Our key idea is to exploit confusi...
Gespeichert in:
Veröffentlicht in: | Machine learning 2020-09, Vol.109 (9-10), p.1837-1853 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Neural networks converge faster with help from a smart batch selection strategy. In this regard, we propose
Ada-Boundary
, a novel and simple adaptive batch selection algorithm that constructs an effective mini-batch according to the learning progress of the model. Our key idea is to exploit
confusing
samples for which the model cannot predict labels with high confidence. Thus, samples near the current decision boundary are considered to be the most effective for expediting convergence. Taking advantage of this design,
Ada-Boundary
maintained its dominance for various degrees of training difficulty. We demonstrate the advantage of
Ada-Boundary
by extensive experimentation using CNNs with five benchmark data sets.
Ada-Boundary
was shown to produce a relative improvement in test errors by up to 31.80% compared with the baseline for a fixed wall-clock training time, thereby achieving a faster convergence speed. |
---|---|
ISSN: | 0885-6125 1573-0565 |
DOI: | 10.1007/s10994-020-05903-6 |