FlatFormer: A Flat Transformer Knowledge Tracing Model Based on Cognitive Bias Injection
By: Xiao-li Xia, Hou-biao Li
Potential Business Impact:
Helps computers track student learning faster.
Knowledge Tracing (KT) models face a critical ``Performance-Complexity Trap'': capturing complex cognitive dynamics like learning sessions and memory decay typically requires deep hierarchical architectures, which incur prohibitive computational costs for real-time deployment. To resolve this, we propose FlatFormer, a streamlined architecture based on the novel design paradigm of ``Information Injection over Structural Stacking.'' Unlike parameter-heavy hierarchical models, FlatFormer leverages a standard flat Transformer augmented with two lightweight injection mechanisms: (i) a hybrid input encoding strategy combining learnable session identifiers with fixed sinusoidal step embeddings; and (ii) a pre-computed power-law bias integrated directly into attention logits to explicitly model the forgetting curve. Extensive experiments on four large-scale datasets (e.g., EdNet, Junyi) show that FlatFormer achieves state-of-the-art performance. For example, on the EdNet dataset, compared to the strongest hierarchical baseline (HiTSKT), its absolute AUC increased by 8.3%, while using less than 15% of parameters, and inference speed was about three times faster. These results validate that high cognitive fidelity does not necessitate architectural complexity.
Similar Papers
ParaFormer: Shallow Parallel Transformers with Progressive Approximation
Machine Learning (CS)
Makes AI models faster and smaller.
UniFormer: Unified and Efficient Transformer for Reasoning Across General and Custom Computing
Distributed, Parallel, and Cluster Computing
Makes AI models work fast on any computer.
Rethinking Transformer Connectivity: TLinFormer, A Path to Exact, Full Context-Aware Linear Attention
Machine Learning (CS)
Makes AI understand long stories faster.