Efficient Knowledge Tracing Leveraging Higher-Order Information in Integrated Graphs
By: Donghee Han , Daehee Kim , Minjun Lee and more
Potential Business Impact:
Makes online learning faster and cheaper.
The rise of online learning has led to the development of various knowledge tracing (KT) methods. However, existing methods have overlooked the problem of increasing computational cost when utilizing large graphs and long learning sequences. To address this issue, we introduce Dual Graph Attention-based Knowledge Tracing (DGAKT), a graph neural network model designed to leverage high-order information from subgraphs representing student-exercise-KC relationships. DGAKT incorporates a subgraph-based approach to enhance computational efficiency. By processing only relevant subgraphs for each target interaction, DGAKT significantly reduces memory and computational requirements compared to full global graph models. Extensive experimental results demonstrate that DGAKT not only outperforms existing KT models but also sets a new standard in resource efficiency, addressing a critical need that has been largely overlooked by prior KT approaches.
Similar Papers
TA-KAND: Two-stage Attention Triple Enhancement and U-KAN based Diffusion For Few-shot Knowledge Graph Completion
Artificial Intelligence
Helps computers learn missing facts from limited examples.
KG-Attention: Knowledge Graph-Guided Attention at Test-Time via Bidirectional Information Aggregation
Computation and Language
Lets computers learn new facts without forgetting old ones.
Extracting Causal Relations in Deep Knowledge Tracing
Artificial Intelligence
Helps computers understand how students learn best.