skip to main content
Caltech

Special CMX Lunch Seminar

Tuesday, December 9, 2025
11:00am to 12:00pm
Add to Cal
Annenberg 105
Understanding Generalization of Deep Generative Models Requires Rethinking Underlying Low-dimensional Structures
Qing Qu, Assistant Professor, Department of Electrical Engineering and Computer Science, University of Michigan,

Diffusion models represent a remarkable new class of deep generative models, yet the mathematical principles underlying their generalization from finite training data are poorly understood. This talk offers novel theoretical insights into diffusion model generalization through the lens of "model reproducibility," revealing a surprising phase transition from memorization to generalization during training, notably occurring without the curse of dimensionality. Our theoretical framework hinges on two crucial observations: (i) the intrinsic low dimensionality of image datasets and (ii) the emergent low-rank property of the denoising autoencoder within trained neural networks. Under simplified settings, we rigorously establish that optimizing the training loss of diffusion models is mathematically equivalent to solving a canonical subspace clustering problem. This insight quantifies the minimal sample requirements for learning low-dimensional distributions, scaling linearly with the intrinsic dimension. Furthermore, by investigating this under a nonlinear two-layer network, we fully explain the memorization-to-generalization transition, highlighting inductive biases in learning dynamics and the models' strong representation learning ability. These theoretical insights have profound practical implications, enabling various applications for generation control and safety, including concept steering, watermarking, and memorization detection. This work not only advances theoretical understanding but also stimulates numerous directions for many applications in engineering and science.

For more information, please contact Jolene Brink by phone at (626)395-2813 or by email at [email protected] or visit CMX Website.