Robust Graph Condensation via Classification Complexity Mitigation
By: Jiayi Luo , Qingyun Sun , Beining Yang and more
Potential Business Impact:
Makes fake graphs work even if real ones are broken.
Graph condensation (GC) has gained significant attention for its ability to synthesize smaller yet informative graphs. However, existing studies often overlook the robustness of GC in scenarios where the original graph is corrupted. In such cases, we observe that the performance of GC deteriorates significantly, while existing robust graph learning technologies offer only limited effectiveness. Through both empirical investigation and theoretical analysis, we reveal that GC is inherently an intrinsic-dimension-reducing process, synthesizing a condensed graph with lower classification complexity. Although this property is critical for effective GC performance, it remains highly vulnerable to adversarial perturbations. To tackle this vulnerability and improve GC robustness, we adopt the geometry perspective of graph data manifold and propose a novel Manifold-constrained Robust Graph Condensation framework named MRGC. Specifically, we introduce three graph data manifold learning modules that guide the condensed graph to lie within a smooth, low-dimensional manifold with minimal class ambiguity, thereby preserving the classification complexity reduction capability of GC and ensuring robust performance under universal adversarial attacks. Extensive experiments demonstrate the robustness of \ModelName\ across diverse attack scenarios.
Similar Papers
Towards Pre-trained Graph Condensation via Optimal Transport
Machine Learning (CS)
Makes computer models learn faster and better.
Multi-view Graph Condensation via Tensor Decomposition
Machine Learning (CS)
Makes big computer graphs smaller for faster learning.
Dynamic Graph Condensation
Machine Learning (CS)
Makes computer learning faster with smaller data.