SFBD Flow: A Continuous-Optimization Framework for Training Diffusion Models with Noisy Samples
By: Haoye Lu, Darren Lo, Yaoliang Yu
Potential Business Impact:
Makes AI create private art without seeing real art.
Diffusion models achieve strong generative performance but often rely on large datasets that may include sensitive content. This challenge is compounded by the models' tendency to memorize training data, raising privacy concerns. SFBD (Lu et al., 2025) addresses this by training on corrupted data and using limited clean samples to capture local structure and improve convergence. However, its iterative denoising and fine-tuning loop requires manual coordination, making it burdensome to implement. We reinterpret SFBD as an alternating projection algorithm and introduce a continuous variant, SFBD flow, that removes the need for alternating steps. We further show its connection to consistency constraint-based methods, and demonstrate that its practical instantiation, Online SFBD, consistently outperforms strong baselines across benchmarks.
Similar Papers
Consistent Flow Distillation for Text-to-3D Generation
CV and Pattern Recognition
Makes 3D pictures from words better.
State Fourier Diffusion Language Model (SFDLM): A Scalable, Novel Iterative Approach to Language Modeling
Machine Learning (CS)
Creates text by fixing jumbled words, no big computers needed.
SoFlow: Solution Flow Models for One-Step Generative Modeling
Machine Learning (CS)
Makes AI create pictures instantly, not slowly.