Score: 1

Diffusion Models and the Manifold Hypothesis: Log-Domain Smoothing is Geometry Adaptive

Published: October 2, 2025 | arXiv ID: 2510.02305v1

By: Tyler Farghly , Peter Potaptchik , Samuel Howard and more

Potential Business Impact:

Teaches AI to learn patterns by smoothing data.

Business Areas:
Semantic Search Internet Services

Diffusion models have achieved state-of-the-art performance, demonstrating remarkable generalisation capabilities across diverse domains. However, the mechanisms underpinning these strong capabilities remain only partially understood. A leading conjecture, based on the manifold hypothesis, attributes this success to their ability to adapt to low-dimensional geometric structure within the data. This work provides evidence for this conjecture, focusing on how such phenomena could result from the formulation of the learning problem through score matching. We inspect the role of implicit regularisation by investigating the effect of smoothing minimisers of the empirical score matching objective. Our theoretical and empirical results confirm that smoothing the score function -- or equivalently, smoothing in the log-density domain -- produces smoothing tangential to the data manifold. In addition, we show that the manifold along which the diffusion model generalises can be controlled by choosing an appropriate smoothing.

Country of Origin
🇬🇧 United Kingdom

Page Count
44 pages

Category
Computer Science:
Machine Learning (CS)