ImageGem: In-the-wild Generative Image Interaction Dataset for Generative Model Personalization
By: Yuanhe Guo , Linxi Xie , Zhuoran Chen and more
Potential Business Impact:
Teaches computers to make pictures you'll love.
We introduce ImageGem, a dataset for studying generative models that understand fine-grained individual preferences. We posit that a key challenge hindering the development of such a generative model is the lack of in-the-wild and fine-grained user preference annotations. Our dataset features real-world interaction data from 57K users, who collectively have built 242K customized LoRAs, written 3M text prompts, and created 5M generated images. With user preference annotations from our dataset, we were able to train better preference alignment models. In addition, leveraging individual user preference, we investigated the performance of retrieval models and a vision-language model on personalized image retrieval and generative model recommendation. Finally, we propose an end-to-end framework for editing customized diffusion models in a latent weight space to align with individual user preferences. Our results demonstrate that the ImageGem dataset enables, for the first time, a new paradigm for generative model personalization.
Similar Papers
DesignPref: Capturing Personal Preferences in Visual Design Generation
CV and Pattern Recognition
Makes AI understand what *you* like in designs.
Learning User Preferences for Image Generation Model
CV and Pattern Recognition
Predicts what you'll like to see next.
GEM-Bench: A Benchmark for Ad-Injected Response Generation within Generative Engine Marketing
Information Retrieval
Makes ads in chatbots better without annoying users.