This paper studies the rates of convergence for learning distributions implicitly with the adversarial framework and Generative Adversarial Networks (GAN), which subsume Wasserstein, Sobolev, MMD GAN, and Generalized/Simulated Method of Moments (GMM/SMM) as special cases. We study a wide range of parametric and nonparametric target distributions, under a host of objective evaluation metrics. We investigate how to obtain a good statistical guarantee for GANs through the lens of regularization. On the nonparametric end, we derive the optimal minimax rates for distribution estimation under the adversarial framework. On the parametric end, we establish a theory for general neural network classes (including deep leaky ReLU networks), that characterizes the interplay on the choice of generator and discriminator pair. We discover and isolate a new notion of regularization, called the generator-discriminator-pair regularization, that sheds light on the advantage of GANs compared to classical parametric and nonparametric approaches for explicit distribution estimation. We develop novel oracle inequalities as the main technical tools for analyzing GANs, which is of independent interest.

More on this topic

BFI Working Paper·Nov 20, 2025

Social Dynamics of AI Adoption

Leonardo Bursztyn, Alex Imas, Rafael Jiménez-Durán, Aaron Leonard, and Christopher Roth
Topics: K-12 Education, Technology & Innovation
BFI Working Paper·Sep 16, 2025

The Promise of Digital Technology and Generative AI for Supporting Parenting Interventions in Latin America

Ariel Kalil, Michelle Michelini, and Pablo Ramos
Topics: Early Childhood Education, Technology & Innovation
BFI Working Paper·Sep 8, 2025

Chat2Learn: A Proof-of-Concept Evaluation of a Technology-Based Tool to Enhance Parent-Child Language Interaction

Linxi Lu and Ariel Kalil
Topics: Early Childhood Education, Technology & Innovation