Score: 0

Iterative Tilting for Diffusion Fine-Tuning

Published: December 2, 2025 | arXiv ID: 2512.03234v1

By: Jean Pachebat , Giovanni Conforti , Alain Durmus and more

Potential Business Impact:

Makes AI create better art by learning from mistakes.

Business Areas:
A/B Testing Data and Analytics

We introduce iterative tilting, a gradient-free method for fine-tuning diffusion models toward reward-tilted distributions. The method decomposes a large reward tilt $\exp(λr)$ into $N$ sequential smaller tilts, each admitting a tractable score update via first-order Taylor expansion. This requires only forward evaluations of the reward function and avoids backpropagating through sampling chains. We validate on a two-dimensional Gaussian mixture with linear reward, where the exact tilted distribution is available in closed form.

Page Count
14 pages

Category
Statistics:
Machine Learning (Stat)