From Prompts to Deployment: Auto-Curated Domain-Specific Dataset Generation via Diffusion Models
By: Dongsik Yoon, Jongeun Kim
In this paper, we present an automated pipeline for generating domain-specific synthetic datasets with diffusion models, addressing the distribution shift between pre-trained models and real-world deployment environments. Our three-stage framework first synthesizes target objects within domain-specific backgrounds through controlled inpainting. The generated outputs are then validated via a multi-modal assessment that integrates object detection, aesthetic scoring, and vision-language alignment. Finally, a user-preference classifier is employed to capture subjective selection criteria. This pipeline enables the efficient construction of high-quality, deployable datasets while reducing reliance on extensive real-world data collection.
Similar Papers
Test-Time Modification: Inverse Domain Transformation for Robust Perception
CV and Pattern Recognition
Makes AI see in new places without retraining.
Semi-Supervised Diversity-Aware Domain Adaptation for 3D Object detection
CV and Pattern Recognition
Teaches self-driving cars to see in new places.
Generative diffusion models for agricultural AI: plant image generation, indoor-to-outdoor translation, and expert preference alignment
CV and Pattern Recognition
Makes AI better at growing crops with fake pictures.