Styleclone: Face Stylization with Diffusion Based Data Augmentation
By: Neeraj Matiyali, Siddharth Srivastava, Gaurav Sharma
Potential Business Impact:
Changes photos to look like a chosen style.
We present StyleClone, a method for training image-to-image translation networks to stylize faces in a specific style, even with limited style images. Our approach leverages textual inversion and diffusion-based guided image generation to augment small style datasets. By systematically generating diverse style samples guided by both the original style images and real face images, we significantly enhance the diversity of the style dataset. Using this augmented dataset, we train fast image-to-image translation networks that outperform diffusion-based methods in speed and quality. Experiments on multiple styles demonstrate that our method improves stylization quality, better preserves source image content, and significantly accelerates inference. Additionally, we provide a systematic evaluation of the augmentation techniques and their impact on stylization performance.
Similar Papers
Leveraging Diffusion Models for Stylization using Multiple Style Images
CV and Pattern Recognition
Changes pictures to look like any art style.
SceneTextStylizer: A Training-Free Scene Text Style Transfer Framework with Diffusion Model
CV and Pattern Recognition
Changes text style in pictures without losing meaning.
DiffStyle360: Diffusion-Based 360° Head Stylization via Style Fusion Attention
CV and Pattern Recognition
Changes 3D heads into any art style instantly.