HiGFA: Hierarchical Guidance for Fine-grained Data Augmentation with Diffusion Models
By: Zhiguang Lu , Qianqian Xu , Peisong Wen and more
Potential Business Impact:
Makes fake pictures better for training AI.
Generative diffusion models show promise for data augmentation. However, applying them to fine-grained tasks presents a significant challenge: ensuring synthetic images accurately capture the subtle, category-defining features critical for high fidelity. Standard approaches, such as text-based Classifier-Free Guidance (CFG), often lack the required specificity, potentially generating misleading examples that degrade fine-grained classifier performance. To address this, we propose Hierarchically Guided Fine-grained Augmentation (HiGFA). HiGFA leverages the temporal dynamics of the diffusion sampling process. It employs strong text and transformed contour guidance with fixed strengths in the early-to-mid sampling stages to establish overall scene, style, and structure. In the final sampling stages, HiGFA activates a specialized fine-grained classifier guidance and dynamically modulates the strength of all guidance signals based on prediction confidence. This hierarchical, confidence-driven orchestration enables HiGFA to generate diverse yet faithful synthetic images by intelligently balancing global structure formation with precise detail refinement. Experiments on several FGVC datasets demonstrate the effectiveness of HiGFA.
Similar Papers
Steering Guidance for Personalized Text-to-Image Diffusion Models
CV and Pattern Recognition
Creates personalized images that match descriptions perfectly.
Dynamic Classifier-Free Diffusion Guidance via Online Feedback
Machine Learning (CS)
Makes AI pictures match words better.
GuidNoise: Single-Pair Guided Diffusion for Generalized Noise Synthesis
CV and Pattern Recognition
Makes blurry photos clear with less work.