Efficient Text-Attributed Graph Learning through Selective Annotation and Graph Alignment
By: Huanyi Xie , Lijie Hu , Lu Yu and more
Potential Business Impact:
Teaches computers to learn from less labeled data.
In the realm of Text-attributed Graphs (TAGs), traditional graph neural networks (GNNs) often fall short due to the complex textual information associated with each node. Recent methods have improved node representations by leveraging large language models (LLMs) to enhance node text features, but these approaches typically require extensive annotations or fine-tuning across all nodes, which is both time-consuming and costly. To overcome these challenges, we introduce GAGA, an efficient framework for TAG representation learning. GAGA reduces annotation time and cost by focusing on annotating only representative nodes and edges. It constructs an annotation graph that captures the topological relationships among these annotations. Furthermore, GAGA employs a two-level alignment module to effectively integrate the annotation graph with the TAG, aligning their underlying structures. Experiments show that GAGA achieves classification accuracies on par with or surpassing state-of-the-art methods while requiring only 1% of the data to be annotated, demonstrating its high efficiency.
Similar Papers
When LLM Agents Meet Graph Optimization: An Automated Data Quality Improvement Approach
Machine Learning (CS)
Fixes messy data for smarter computer graphs.
Navigating the Black Box: Leveraging LLMs for Effective Text-Level Graph Injection Attacks
Artificial Intelligence
Makes fake text fool smart computer graphs.
An Effective Approach for Node Classification in Textual Graphs
Machine Learning (CS)
Helps computers understand research papers better.