Data and Channel-Adaptive Sensor Scheduling for Federated Edge Learning via Over-the-Air Gradient Aggregation

Over-the-air gradient aggregation and data-aware scheduling have recently drawn great attention due to the outstanding performance in improving communication efficiency for federated edge learning applications. However, in this case, the estimated gradient suffers from the channel and data distortio...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE internet of things journal 2022-02, Vol.9 (3), p.1640-1654
Hauptverfasser: Su, Liqun, Lau, Vincent K. N.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Over-the-air gradient aggregation and data-aware scheduling have recently drawn great attention due to the outstanding performance in improving communication efficiency for federated edge learning applications. However, in this case, the estimated gradient suffers from the channel and data distortion induced by channel fading and data-aware scheduling, which introduces significant bias and harms the training performance. To solve these problems, we propose a dynamic data and channel adaptive sensor scheduling and power control algorithm combining a residual feedback mechanism. Instead of discarding the gradients not transmitted to the central server, each sensor keeps track of a local residual to store these gradients. Furthermore, by connecting the model update iterations to a dynamic evolution process, we utilize the Lyapunov drift optimization method to analyze the relationship between the training gain and resource allocation. The derived decentralized optimal solution is adaptive to both the channel state information and data importance to seize good transmission opportunity and important gradients. Theoretical analysis is provided on the convergence of the proposed algorithm in practical training scenarios. Simulation results further illustrate that under the same power cost, the proposed scheme has a much faster convergence rate and lower training loss compared to existing baselines.
ISSN:2327-4662
2327-4662
DOI:10.1109/JIOT.2021.3096570