TY - GEN
T1 - Local Convergence of Gradient Descent-Ascent for Training Generative Adversarial Networks
AU - Becker, Evan
AU - Pandit, Parthe
AU - Rangan, Sundeep
AU - Fletcher, Alyson K.
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Generative Adversarial Networks (GANs) are a popular formulation to train generative models for complex high dimensional data. The standard method for training GANs involves a gradient descent-ascent (GDA) procedure on a minimax optimization problem. This procedure is hard to analyze in general due to the nonlinear nature of the dynamics. We study the local dynamics of GDA for training a GAN with a kernel-based discriminator. This convergence analysis is based on a linearization of a nonlinear dynamical system that describes the GDA iterations, under an isolated points model assumption from [2]. Our analysis brings out the effect of the learning rates, regularization, and the bandwidth of the kernel discriminator, on the local convergence rate of GDA. Importantly, we show phase transitions that indicate when the system converges, oscillates, or diverges. We also provide numerical simulations that verify our claims. A full version with complete proofs is available on arXiv [3].
AB - Generative Adversarial Networks (GANs) are a popular formulation to train generative models for complex high dimensional data. The standard method for training GANs involves a gradient descent-ascent (GDA) procedure on a minimax optimization problem. This procedure is hard to analyze in general due to the nonlinear nature of the dynamics. We study the local dynamics of GDA for training a GAN with a kernel-based discriminator. This convergence analysis is based on a linearization of a nonlinear dynamical system that describes the GDA iterations, under an isolated points model assumption from [2]. Our analysis brings out the effect of the learning rates, regularization, and the bandwidth of the kernel discriminator, on the local convergence rate of GDA. Importantly, we show phase transitions that indicate when the system converges, oscillates, or diverges. We also provide numerical simulations that verify our claims. A full version with complete proofs is available on arXiv [3].
UR - http://www.scopus.com/inward/record.url?scp=85190389847&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85190389847&partnerID=8YFLogxK
U2 - 10.1109/IEEECONF59524.2023.10476957
DO - 10.1109/IEEECONF59524.2023.10476957
M3 - Conference contribution
AN - SCOPUS:85190389847
T3 - Conference Record - Asilomar Conference on Signals, Systems and Computers
SP - 892
EP - 896
BT - Conference Record of the 57th Asilomar Conference on Signals, Systems and Computers, ACSSC 2023
A2 - Matthews, Michael B.
PB - IEEE Computer Society
T2 - 57th Asilomar Conference on Signals, Systems and Computers, ACSSC 2023
Y2 - 29 October 2023 through 1 November 2023
ER -