Enhancing Convergence of Decentralized Gradient Tracking under the KL Property
We study decentralized multiagent optimization over networks, modeled as undirected graphs. The optimization problem consists of minimizing a nonconvex smooth function plus a convex extended-value function, which enforces constraints or extra structure on the solution (e.g., sparsity, low-rank). We...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We study decentralized multiagent optimization over networks, modeled as
undirected graphs. The optimization problem consists of minimizing a nonconvex
smooth function plus a convex extended-value function, which enforces
constraints or extra structure on the solution (e.g., sparsity, low-rank). We
further assume that the objective function satisfies the Kurdyka-{\L}ojasiewicz
(KL) property, with given exponent $\theta\in [0,1)$. The KL property is
satisfied by several (nonconvex) functions of practical interest, e.g., arising
from machine learning applications; in the centralized setting, it permits to
achieve strong convergence guarantees. Here we establish convergence of the
same type for the notorious decentralized gradient-tracking-based algorithm
SONATA. Specifically, $\textbf{(i)}$ when $\theta\in (0,1/2]$, the sequence
generated by SONATA converges to a stationary solution of the problem at
R-linear rate;$ \textbf{(ii)} $when $\theta\in (1/2,1)$, sublinear rate is
certified; and finally $\textbf{(iii)}$ when $\theta=0$, the iterates will
either converge in a finite number of steps or converges at R-linear rate. This
matches the convergence behavior of centralized proximal-gradient algorithms
except when $\theta=0$. Numerical results validate our theoretical findings. |
---|---|
DOI: | 10.48550/arxiv.2412.09556 |