On the convergence of single-call stochastic extra-gradient methods
Variational inequalities have recently attracted considerable interest in machine learning as
a flexible paradigm for models that go beyond ordinary loss function minimization (such as …
a flexible paradigm for models that go beyond ordinary loss function minimization (such as …
Distributed variable sample-size gradient-response and best-response schemes for stochastic Nash equilibrium problems
J Lei, UV Shanbhag - SIAM Journal on Optimization, 2022 - SIAM
This paper considers an n-player stochastic Nash equilibrium problem (NEP) in which the i
th player minimizes a composite objective f_i(∙,x_-i)+r_i(∙), where f_i is an expectation-valued …
th player minimizes a composite objective f_i(∙,x_-i)+r_i(∙), where f_i is an expectation-valued …
Two Steps at a Time---Taking GAN Training in Stride with Tseng's Method
Motivated by the training of generative adversarial networks (GANs), we study methods for
solving minimax problems with additional nonsmooth regularizers. We do so by employing …
solving minimax problems with additional nonsmooth regularizers. We do so by employing …
An optimal multistage stochastic gradient method for minimax problems
In this paper, we study the minimax optimization problem in the smooth and strongly convex-
strongly concave setting when we have access to noisy estimates of gradients. In particular …
strongly concave setting when we have access to noisy estimates of gradients. In particular …
Forward-reflected-backward method with variance reduction
We propose a variance reduced algorithm for solving monotone variational inequalities.
Without assuming strong monotonicity, cocoercivity, or boundedness of the domain, we …
Without assuming strong monotonicity, cocoercivity, or boundedness of the domain, we …
A distributed forward–backward algorithm for stochastic generalized Nash equilibrium seeking
B Franci, S Grammatico - IEEE Transactions on Automatic …, 2020 - ieeexplore.ieee.org
We consider the stochastic generalized Nash equilibrium problem (SGNEP) with expected-
value cost functions. Inspired by Yi and Pavel (2019), we propose a distributed generalized …
value cost functions. Inspired by Yi and Pavel (2019), we propose a distributed generalized …
On the convergence of stochastic extragradient for bilinear games using restarted iteration averaging
We study the stochastic bilinear minimax optimization problem, presenting an analysis of the
same-sample Stochastic ExtraGradient (SEG) method with constant step size, and …
same-sample Stochastic ExtraGradient (SEG) method with constant step size, and …
Stochastic generalized Nash equilibrium-seeking in merely monotone games
B Franci, S Grammatico - IEEE Transactions on Automatic …, 2021 - ieeexplore.ieee.org
We solve the stochastic generalized Nash equilibrium (SGNE) problem in merely monotone
games with expected value cost functions. Specifically, we present the first distributed SGNE …
games with expected value cost functions. Specifically, we present the first distributed SGNE …
Stochastic variance-reduced forward-reflected methods for root-finding problems
Q Tran-Dinh - arXiv preprint arXiv:2406.00937, 2024 - arxiv.org
We develop two novel stochastic variance-reduction methods to approximate a solution of
root-finding problems applicable to both equations and inclusions. Our algorithms leverage …
root-finding problems applicable to both equations and inclusions. Our algorithms leverage …
Training generative adversarial networks via stochastic Nash games
B Franci, S Grammatico - IEEE Transactions on Neural …, 2021 - ieeexplore.ieee.org
Generative adversarial networks (GANs) are a class of generative models with two
antagonistic neural networks: a generator and a discriminator. These two neural networks …
antagonistic neural networks: a generator and a discriminator. These two neural networks …