The Randomized Block Coordinate Descent Method in the H\"older Smooth Setting

This work provides the first convergence analysis for the Randomized Block Coordinate Descent method for minimizing a function that is both H\"older smooth and block H\"older smooth. Our analysis applies to objective functions that are non-convex, convex, and strongly convex. For non-conve...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Maia, Leandro Farias, Gutman, David Huckleberry
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:This work provides the first convergence analysis for the Randomized Block Coordinate Descent method for minimizing a function that is both H\"older smooth and block H\"older smooth. Our analysis applies to objective functions that are non-convex, convex, and strongly convex. For non-convex functions, we show that the expected gradient norm reduces at an $O\left(k^{\frac{\gamma}{1+\gamma}}\right)$ rate, where $k$ is the iteration count and $\gamma$ is the H\"older exponent. For convex functions, we show that the expected suboptimality gap reduces at the rate $O\left(k^{-\gamma}\right)$. In the strongly convex setting, we show this rate for the expected suboptimality gap improves to $O\left(k^{-\frac{2\gamma}{1-\gamma}}\right)$ when $\gamma>1$ and to a linear rate when $\gamma=1$. Notably, these new convergence rates coincide with those furnished in the existing literature for the Lipschitz smooth setting.
DOI:10.48550/arxiv.2403.08080