ControlFill: Spatially Adjustable Image Inpainting from Prompt Learning
By: Boseong Jeon
Potential Business Impact:
Lets you add or remove things from pictures easily.
In this report, I present an inpainting framework named \textit{ControlFill}, which involves training two distinct prompts: one for generating plausible objects within a designated mask (\textit{creation}) and another for filling the region by extending the background (\textit{removal}). During the inference stage, these learned embeddings guide a diffusion network that operates without requiring heavy text encoders. By adjusting the relative significance of the two prompts and employing classifier-free guidance, users can control the intensity of removal or creation. Furthermore, I introduce a method to spatially vary the intensity of guidance by assigning different scales to individual pixels.
Similar Papers
FreeInpaint: Tuning-free Prompt Alignment and Visual Rationality Enhancement in Image Inpainting
CV and Pattern Recognition
Makes AI fill in missing picture parts from your words.
CorrFill: Enhancing Faithfulness in Reference-based Inpainting with Correspondence Guidance in Diffusion Models
CV and Pattern Recognition
Fixes damaged pictures using a matching photo.
PixelHacker: Image Inpainting with Structural and Semantic Consistency
CV and Pattern Recognition
Fixes missing parts of pictures perfectly.