Task-Aware Multi-Expert Architecture For Lifelong Deep Learning
By: Jianyu Wang , Jacob Nean-Hua Sheikh , Cat P. Le and more
Lifelong deep learning (LDL) trains neural networks to learn sequentially across tasks while preserving prior knowledge. We propose Task-Aware Multi-Expert (TAME), a continual learning algorithm that leverages task similarity to guide expert selection and knowledge transfer. TAME maintains a pool of pretrained neural networks and activates the most relevant expert for each new task. A shared dense layer integrates features from the chosen expert to generate predictions. To reduce catastrophic forgetting, TAME uses a replay buffer that stores representative samples and embeddings from previous tasks and reuses them during training. An attention mechanism further prioritizes the most relevant stored information for each prediction. Together, these components allow TAME to adapt flexibly while retaining important knowledge across evolving task sequences. Experiments on binary classification tasks derived from CIFAR-100 show that TAME improves accuracy on new tasks while sustaining performance on earlier ones, highlighting its effectiveness in balancing adaptation and retention in lifelong learning settings.
Similar Papers
Dynamic Mixture of Experts Against Severe Distribution Shifts
Machine Learning (CS)
Lets computers learn new things without forgetting old ones.
Bridging Lifelong and Multi-Task Representation Learning via Algorithm and Complexity Measure
Machine Learning (CS)
Teaches computers to learn new things faster.
Drift-aware Collaborative Assistance Mixture of Experts for Heterogeneous Multistream Learning
Machine Learning (CS)
Helps computers learn from many different, changing information sources.