Distributed Adaptive Subgradient Algorithms for Online Learning Over Time-Varying Networks

Adaptive gradient algorithms have recently become extremely popular because they have been applied successfully in training deep neural networks, such as Adam, AMSGrad, and AdaBound. Despite their success, however, the distributed variant of the adaptive method, which is expected to possess a rapid...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on systems, man, and cybernetics. Systems man, and cybernetics. Systems, 2022-07, Vol.52 (7), p.1-12
Hauptverfasser: Zhang, Mingchuan, Hao, Bowei, Ge, Quanbo, Zhu, Junlong, Zheng, Ruijuan, Wu, Qingtao
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Adaptive gradient algorithms have recently become extremely popular because they have been applied successfully in training deep neural networks, such as Adam, AMSGrad, and AdaBound. Despite their success, however, the distributed variant of the adaptive method, which is expected to possess a rapid training speed at the beginning and a good generalization capacity at the end, is rarely studied. To fill the gap, a distributed adaptive subgradient algorithm is presented, called D-AdaBound, where the learning rates are dynamically bounded by clipping the learning rates. Moreover, we obtain the regret bound of D-AdaBound, in which the objective functions are convex. Finally, we confirm the effectiveness of D-AdaBound by simulation experiments on different datasets. The results show the performance improvement of D-AdaBound relative to existing distributed online learning algorithms.
ISSN:2168-2216
2168-2232
DOI:10.1109/TSMC.2021.3097714