Variance-Based Single-Call Proximal Extragradient Algorithms for Stochastic Mixed Variational Inequalities
In the study of stochastic variational inequalities, the extragradient algorithms attract much attention. However, such schemes require two evaluations of the expected mapping at each iteration in general. In this paper, we present several variance-based single-call proximal extragradient algorithms...
Gespeichert in:
Veröffentlicht in: | Journal of optimization theory and applications 2021-08, Vol.190 (2), p.393-427 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | In the study of stochastic variational inequalities, the extragradient algorithms attract much attention. However, such schemes require two evaluations of the expected mapping at each iteration in general. In this paper, we present several variance-based single-call proximal extragradient algorithms for solving a class of stochastic mixed variational inequalities by aiming at alleviating the cost of an extragradient step. One salient feature of the proposed algorithms is that they require only one evaluation of the expected mapping at each iteration, and hence, the computation load may be significantly reduced. We show that the proposed algorithms can achieve sublinear ergodic convergence rate in terms of the restricted merit function. Furthermore, under the strongly Minty variational inequality condition, we derive some results related to convergence rate of the distance between iterates and solutions, the iteration and oracle complexities for the proposed algorithms when the sample size increases at a geometric or polynomial rate. Numerical experiments indicate that the proposed algorithms are quite competitive with some existing algorithms. |
---|---|
ISSN: | 0022-3239 1573-2878 |
DOI: | 10.1007/s10957-021-01882-3 |