On Arbitrary Compression for Decentralized Consensus and Stochastic Optimization over Directed Networks
We study the decentralized consensus and stochastic optimization problems with compressed communications over static directed graphs. We propose an iterative gradient-based algorithm that compresses messages according to a desired compression ratio. The proposed method provably reduces the communica...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We study the decentralized consensus and stochastic optimization problems
with compressed communications over static directed graphs. We propose an
iterative gradient-based algorithm that compresses messages according to a
desired compression ratio. The proposed method provably reduces the
communication overhead on the network at every communication round. Contrary to
existing literature, we allow for arbitrary compression ratios in the
communicated messages. We show a linear convergence rate for the proposed
method on the consensus problem. Moreover, we provide explicit convergence
rates for decentralized stochastic optimization problems on smooth functions
that are either (i) strongly convex, (ii) convex, or (iii) non-convex. Finally,
we provide numerical experiments to illustrate convergence under arbitrary
compression ratios and the communication efficiency of our algorithm. |
---|---|
DOI: | 10.48550/arxiv.2204.08160 |