Invariant Graph Transformer for Out-of-Distribution Generalization
By: Tianyin Liao , Ziwei Zhang , Yufei Sun and more
Potential Business Impact:
Helps computers understand different kinds of graphs.
Graph Transformers (GTs) have demonstrated great effectiveness across various graph analytical tasks. However, the existing GTs focus on training and testing graph data originated from the same distribution, but fail to generalize under distribution shifts. Graph invariant learning, aiming to capture generalizable graph structural patterns with labels under distribution shifts, is potentially a promising solution, but how to design attention mechanisms and positional and structural encodings (PSEs) based on graph invariant learning principles remains challenging. To solve these challenges, we introduce Graph Out-Of-Distribution generalized Transformer (GOODFormer), aiming to learn generalized graph representations by capturing invariant relationships between predictive graph structures and labels through jointly optimizing three modules. Specifically, we first develop a GT-based entropy-guided invariant subgraph disentangler to separate invariant and variant subgraphs while preserving the sharpness of the attention function. Next, we design an evolving subgraph positional and structural encoder to effectively and efficiently capture the encoding information of dynamically changing subgraphs during training. Finally, we propose an invariant learning module utilizing subgraph node representations and encodings to derive generalizable graph representations that can to unseen graphs. We also provide theoretical justifications for our method. Extensive experiments on benchmark datasets demonstrate the superiority of our method over state-of-the-art baselines under distribution shifts.
Similar Papers
Evolving Graph Learning for Out-of-Distribution Generalization in Non-stationary Environments
Machine Learning (CS)
Helps computers learn from changing data better.
Fully Inductive Node Representation Learning via Graph View Transformation
Machine Learning (CS)
Lets computers learn from new graph data.
Quantifying Distributional Invariance in Causal Subgraph for IRM-Free Graph Generalization
Machine Learning (CS)
Finds important graph parts that work everywhere.