Bits for Privacy: Evaluating Post-Training Quantization via Membership Inference
By: Chenxiang Zhang , Tongxi Qu , Zhong Li and more
Potential Business Impact:
Makes AI models more private by using less detail.
Deep neural networks are widely deployed with quantization techniques to reduce memory and computational costs by lowering the numerical precision of their parameters. While quantization alters model parameters and their outputs, existing privacy analyses primarily focus on full-precision models, leaving a gap in understanding how bit-width reduction can affect privacy leakage. We present the first systematic study of the privacy-utility relationship in post-training quantization (PTQ), a versatile family of methods that can be applied to pretrained models without further training. Using membership inference attacks as our evaluation framework, we analyze three popular PTQ algorithms-AdaRound, BRECQ, and OBC-across multiple precision levels (4-bit, 2-bit, and 1.58-bit) on CIFAR-10, CIFAR-100, and TinyImageNet datasets. Our findings consistently show that low-precision PTQs can reduce privacy leakage. In particular, lower-precision models demonstrate up to an order of magnitude reduction in membership inference vulnerability compared to their full-precision counterparts, albeit at the cost of decreased utility. Additional ablation studies on the 1.58-bit quantization level show that quantizing only the last layer at higher precision enables fine-grained control over the privacy-utility trade-off. These results offer actionable insights for practitioners to balance efficiency, utility, and privacy protection in real-world deployments.
Similar Papers
Sensitivity-Aware Post-Training Quantization for Deep Neural Networks
CV and Pattern Recognition
Makes smart computer programs smaller, faster, and still accurate.
A Quantized VAE-MLP Botnet Detection Model: A Systematic Evaluation of Quantization-Aware Training and Post-Training Quantization Strategies
Machine Learning (CS)
Makes small devices catch internet hackers faster.
Privacy-Preserving Inference for Quantized BERT Models
Machine Learning (CS)
Keeps your private data safe during AI use.