On the Generalization of Wasserstein Robust Federated Learning
In federated learning, participating clients typically possess non-i.i.d. data, posing a significant challenge to generalization to unseen distributions. To address this, we propose a Wasserstein distributionally robust optimization scheme called WAFL. Leveraging its duality, we frame WAFL as an emp...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | In federated learning, participating clients typically possess non-i.i.d.
data, posing a significant challenge to generalization to unseen distributions.
To address this, we propose a Wasserstein distributionally robust optimization
scheme called WAFL. Leveraging its duality, we frame WAFL as an empirical
surrogate risk minimization problem, and solve it using a local SGD-based
algorithm with convergence guarantees. We show that the robustness of WAFL is
more general than related approaches, and the generalization bound is robust to
all adversarial distributions inside the Wasserstein ball (ambiguity set).
Since the center location and radius of the Wasserstein ball can be suitably
modified, WAFL shows its applicability not only in robustness but also in
domain adaptation. Through empirical evaluation, we demonstrate that WAFL
generalizes better than the vanilla FedAvg in non-i.i.d. settings, and is more
robust than other related methods in distribution shift settings. Further,
using benchmark datasets we show that WAFL is capable of generalizing to unseen
target domains. |
---|---|
DOI: | 10.48550/arxiv.2206.01432 |