EcoSta 2018

Abstract

We discuss the statistical and computational theory for Generative Adversarial Networks (GANs) in this talk. On the statistical side, we study the rate of convergence for learning densities under the GANs framework, borrowing insights from nonparametric statistics. We introduce an improved GAN estimator that achieves a faster rate, through leveraging the level of smoothness in the target density and the evaluation metric, which in theory remedies the mode collapse problem reported in the literature. A minimax lower bound is constructed to show that when the dimension is large, the exponent in the rate for the new GAN estimator is near optimal. As a byproduct, we also obtain improved bounds for GAN with deeper ReLU discriminator network. On the computational and algorithmic side, we present a simple yet unified non-asymptotic local convergence theory for smooth two-player games, which subsumes several discrete-time saddle point dynamics. The analysis reveals the surprising nature of the off-diagonal interaction term as both a blessing and a curse. On the one hand, this interaction term explains the origin of the slow-down effect in the convergence of Simultaneous Gradient Ascent (SGA) to stable Nash equilibria. On the other hand, for the unstable equilibria, exponential convergence can be proved thanks to the interaction term, for three modified dynamics which have been proposed to stabilize GAN training: Optimistic Mirror Descent (OMD), Consensus Optimization (CO) and Predictive Method (PM). The analysis uncovers the intimate connections among these stabilizing techniques, and provides detailed characterization on the choice of learning rate.

Date
Event
Frontiers in Financial Statistics Invited Session
Location
LT-13