Controllable Coupled Image Generation via Diffusion Models
By: Chenfei Yuan , Nanshan Jia , Hangqi Li and more
Potential Business Impact:
Creates many pictures with same background, different objects.
We provide an attention-level control method for the task of coupled image generation, where "coupled" means that multiple simultaneously generated images are expected to have the same or very similar backgrounds. While backgrounds coupled, the centered objects in the generated images are still expected to enjoy the flexibility raised from different text prompts. The proposed method disentangles the background and entity components in the model's cross-attention modules, attached with a sequence of time-varying weight control parameters depending on the time step of sampling. We optimize this sequence of weight control parameters with a combined objective that assesses how coupled the backgrounds are as well as text-to-image alignment and overall visual quality. Empirical results demonstrate that our method outperforms existing approaches across these criteria.
Similar Papers
Bokeh Diffusion: Defocus Blur Control in Text-to-Image Diffusion Models
Graphics
Makes AI pictures look like real photos.
Controllable Motion Generation via Diffusion Modal Coupling
Robotics
Robots can now choose the best way to move.
Multi-party Collaborative Attention Control for Image Customization
CV and Pattern Recognition
Makes AI create custom pictures from text and images.