OpenLVLM-MIA: A Controlled Benchmark Revealing the Limits of Membership Inference Attacks on Large Vision-Language Models
By: Ryoto Miyamoto , Xin Fan , Fuyuko Kido and more
Potential Business Impact:
Finds if AI remembers private training pictures.
OpenLVLM-MIA is a new benchmark that highlights fundamental challenges in evaluating membership inference attacks (MIA) against large vision-language models (LVLMs). While prior work has reported high attack success rates, our analysis suggests that these results often arise from detecting distributional bias introduced during dataset construction rather than from identifying true membership status. To address this issue, we introduce a controlled benchmark of 6{,}000 images where the distributions of member and non-member samples are carefully balanced, and ground-truth membership labels are provided across three distinct training stages. Experiments using OpenLVLM-MIA demonstrated that the performance of state-of-the-art MIA methods converged to random chance under unbiased conditions. By offering a transparent and unbiased benchmark, OpenLVLM-MIA clarifies the current limitations of MIA research on LVLMs and provides a solid foundation for developing stronger privacy-preserving techniques.
Similar Papers
Image Corruption-Inspired Membership Inference Attacks against Large Vision-Language Models
CV and Pattern Recognition
Finds if your pictures trained AI.
Black-Box Membership Inference Attack for LVLMs via Prior Knowledge-Calibrated Memory Probing
Cryptography and Security
Uncovers if AI saw your private pictures.
Revisiting Data Auditing in Large Vision-Language Models
CV and Pattern Recognition
Finds if AI saw your private pictures.