Stealthy Yet Effective: Distribution-Preserving Backdoor Attacks on Graph Classification
By: Xiaobao Wang , Ruoxiao Sun , Yujun Zhang and more
Potential Business Impact:
Hides secret messages in computer networks.
Graph Neural Networks (GNNs) have demonstrated strong performance across tasks such as node classification, link prediction, and graph classification, but remain vulnerable to backdoor attacks that implant imperceptible triggers during training to control predictions. While node-level attacks exploit local message passing, graph-level attacks face the harder challenge of manipulating global representations while maintaining stealth. We identify two main sources of anomaly in existing graph classification backdoor methods: structural deviation from rare subgraph triggers and semantic deviation caused by label flipping, both of which make poisoned graphs easily detectable by anomaly detection models. To address this, we propose DPSBA, a clean-label backdoor framework that learns in-distribution triggers via adversarial training guided by anomaly-aware discriminators. DPSBA effectively suppresses both structural and semantic anomalies, achieving high attack success while significantly improving stealth. Extensive experiments on real-world datasets validate that DPSBA achieves a superior balance between effectiveness and detectability compared to state-of-the-art baselines.
Similar Papers
A Semantic and Clean-label Backdoor Attack against Graph Convolutional Networks
Machine Learning (CS)
Makes smart computer programs easily tricked.
Cross-Paradigm Graph Backdoor Attacks with Promptable Subgraph Triggers
Cryptography and Security
Makes computer attacks work on different systems.
Towards Effective, Stealthy, and Persistent Backdoor Attacks Targeting Graph Foundation Models
Cryptography and Security
Makes smart computer graphs secretly do wrong things.