Identifying Models Behind Text-to-Image Leaderboards
By: Ali Naseh , Yuefeng Peng , Anshuman Suri and more
Potential Business Impact:
Identifies which AI made a picture.
Text-to-image (T2I) models are increasingly popular, producing a large share of AI-generated images online. To compare model quality, voting-based leaderboards have become the standard, relying on anonymized model outputs for fairness. In this work, we show that such anonymity can be easily broken. We find that generations from each T2I model form distinctive clusters in the image embedding space, enabling accurate deanonymization without prompt control or training data. Using 22 models and 280 prompts (150K images), our centroid-based method achieves high accuracy and reveals systematic model-specific signatures. We further introduce a prompt-level distinguishability metric and conduct large-scale analyses showing how certain prompts can lead to near-perfect distinguishability. Our findings expose fundamental security flaws in T2I leaderboards and motivate stronger anonymization defenses.
Similar Papers
Text-to-Image Models Leave Identifiable Signatures: Implications for Leaderboard Security
Machine Learning (CS)
Makes AI art generators easier to trick.
T2IBias: Uncovering Societal Bias Encoded in the Latent Space of Text-to-Image Generative Models
Machine Learning (CS)
AI makes pictures show unfair stereotypes.
Hierarchical Prompt Learning for Image- and Text-Based Person Re-Identification
CV and Pattern Recognition
Find people in photos using pictures or words.