1. Memorization and Generalization in Generative Diffusion under the Manifold Hypothesis
- Author
-
Achilli, Beatrice, Ambrogioni, Luca, Lucibello, Carlo, Mézard, Marc, and Ventura, Enrico
- Subjects
Condensed Matter - Disordered Systems and Neural Networks - Abstract
We study the memorization and generalization capabilities of a Diffusion Model (DM) in the case of structured data defined on a latent manifold. We specifically consider a set of $P$ mono-modal data points in $N$ dimensions lying on a latent subspace of dimension $D = \alpha_D N$, according to the Hidden Manifold Model (HMM). Our analysis leverages the recently introduced formalism based on the statistical physics of the Random Energy Model (REM). We provide evidence for the existence of an onset time $t_{o} > t_c$ when traps appear in the potential without affecting the typical diffusive trajectory. The size of the basins of attraction of such traps is computed as a function of time. Moreover, we derive the collapse time $t_{c}$ at which trajectories fall in the basin of one of the training points, implying memorization. An explicit formula for $t_c$ is given as a function of $P$ and the ratio $\alpha_D$, proving that the curse of dimensionality issue does not hold for highly structured data, i.e. $\alpha_D\ll 1$, regardless of the non-linearity of the manifold surface. We also prove that collapse coincides with the condensation transition in the REM. Eventually, the degree of generalization of DMs is formulated in terms of the Kullback-Leibler divergence between the exact and the empirical distribution of the sampled configurations: we show the existence of an additional time $t_{g}
- Published
- 2025