Variance-Based Single-Call Proximal Extragradient Algorithms for Stochastic Mixed Variational Inequalities

In the study of stochastic variational inequalities, the extragradient algorithms attract much attention. However, such schemes require two evaluations of the expected mapping at each iteration in general. In this paper, we present several variance-based single-call proximal extragradient algorithms...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of optimization theory and applications 2021-08, Vol.190 (2), p.393-427
Hauptverfasser: Yang, Zhen-Ping, Lin, Gui-Hua
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:In the study of stochastic variational inequalities, the extragradient algorithms attract much attention. However, such schemes require two evaluations of the expected mapping at each iteration in general. In this paper, we present several variance-based single-call proximal extragradient algorithms for solving a class of stochastic mixed variational inequalities by aiming at alleviating the cost of an extragradient step. One salient feature of the proposed algorithms is that they require only one evaluation of the expected mapping at each iteration, and hence, the computation load may be significantly reduced. We show that the proposed algorithms can achieve sublinear ergodic convergence rate in terms of the restricted merit function. Furthermore, under the strongly Minty variational inequality condition, we derive some results related to convergence rate of the distance between iterates and solutions, the iteration and oracle complexities for the proposed algorithms when the sample size increases at a geometric or polynomial rate. Numerical experiments indicate that the proposed algorithms are quite competitive with some existing algorithms.
ISSN:0022-3239
1573-2878
DOI:10.1007/s10957-021-01882-3