Salient Concept-Aware Generative Data Augmentation
By: Tianchen Zhao , Xuanbai Chen , Zhihua Li and more
Potential Business Impact:
Makes AI create better, more varied pictures from words.
Recent generative data augmentation methods conditioned on both image and text prompts struggle to balance between fidelity and diversity, as it is challenging to preserve essential image details while aligning with varied text prompts. This challenge arises because representations in the synthesis process often become entangled with non-essential input image attributes such as environmental contexts, creating conflicts with text prompts intended to modify these elements. To address this, we propose a personalized image generation framework that uses a salient concept-aware image embedding model to reduce the influence of irrelevant visual details during the synthesis process, thereby maintaining intuitive alignment between image and text inputs. By generating images that better preserve class-discriminative features with additional controlled variations, our framework effectively enhances the diversity of training datasets and thereby improves the robustness of downstream models. Our approach demonstrates superior performance across eight fine-grained vision datasets, outperforming state-of-the-art augmentation methods with averaged classification accuracy improvements by 0.73% and 6.5% under conventional and long-tail settings, respectively.
Similar Papers
Semantic Anchoring for Robust Personalization in Text-to-Image Diffusion Models
CV and Pattern Recognition
Teaches AI to draw your specific things from photos.
Generative Hints
CV and Pattern Recognition
Teaches computers to see things better.
Per-Query Visual Concept Learning
CV and Pattern Recognition
Teaches computers to draw your specific ideas.