Loss Functions in Diffusion Models: A Comparative Study
By: Dibyanshu Kumar, Philipp Vaeth, Magda Gregorová
Potential Business Impact:
Makes AI create better pictures and understand them.
Diffusion models have emerged as powerful generative models, inspiring extensive research into their underlying mechanisms. One of the key questions in this area is the loss functions these models shall train with. Multiple formulations have been introduced in the literature over the past several years with some links and some critical differences stemming from various initial considerations. In this paper, we explore the different target objectives and corresponding loss functions in detail. We present a systematic overview of their relationships, unifying them under the framework of the variational lower bound objective. We complement this theoretical analysis with an empirical study providing insights into the conditions under which these objectives diverge in performance and the underlying factors contributing to such deviations. Additionally, we evaluate how the choice of objective impacts the model ability to achieve specific goals, such as generating high-quality samples or accurately estimating likelihoods. This study offers a unified understanding of loss functions in diffusion models, contributing to more efficient and goal-oriented model designs in future research.
Similar Papers
Diagnosing and Improving Diffusion Models by Estimating the Optimal Loss Value
Machine Learning (CS)
Helps AI make better pictures by checking its work.
Task-based Loss Functions in Computer Vision: A Comprehensive Review
Machine Learning (CS)
Teaches computers to learn better from mistakes.
Rethinking Diffusion Model in High Dimension
Machine Learning (Stat)
Makes computers create realistic pictures from simple ideas.