Attention Beyond Neighborhoods: Reviving Transformer for Graph Clustering
By: Xuanting Xie , Bingheng Li , Erlin Pan and more
Potential Business Impact:
Helps computers group similar things by looking at connections.
Attention mechanisms have become a cornerstone in modern neural networks, driving breakthroughs across diverse domains. However, their application to graph structured data, where capturing topological connections is essential, remains underexplored and underperforming compared to Graph Neural Networks (GNNs), particularly in the graph clustering task. GNN tends to overemphasize neighborhood aggregation, leading to a homogenization of node representations. Conversely, Transformer tends to over globalize, highlighting distant nodes at the expense of meaningful local patterns. This dichotomy raises a key question: Is attention inherently redundant for unsupervised graph learning? To address this, we conduct a comprehensive empirical analysis, uncovering the complementary weaknesses of GNN and Transformer in graph clustering. Motivated by these insights, we propose the Attentive Graph Clustering Network (AGCN) a novel architecture that reinterprets the notion that graph is attention. AGCN directly embeds the attention mechanism into the graph structure, enabling effective global information extraction while maintaining sensitivity to local topological cues. Our framework incorporates theoretical analysis to contrast AGCN behavior with GNN and Transformer and introduces two innovations: (1) a KV cache mechanism to improve computational efficiency, and (2) a pairwise margin contrastive loss to boost the discriminative capacity of the attention space. Extensive experimental results demonstrate that AGCN outperforms state-of-the-art methods.
Similar Papers
When Does Global Attention Help? A Unified Empirical Study on Atomistic Graph Learning
Machine Learning (CS)
Helps computers predict material properties faster.
Exploring the Global-to-Local Attention Scheme in Graph Transformers: An Empirical Study
Machine Learning (CS)
Helps computers understand complex connections better.
Topologic Attention Networks: Attending to Direct and Indirect Neighbors through Gaussian Belief Propagation
Machine Learning (CS)
Lets computers understand complex connections faster.