Multi-level Collaborative Distillation Meets Global Workspace Model: A Unified Framework for OCIL
By: Shibin Su , Guoqiang Liang , De Cheng and more
Potential Business Impact:
Helps computers learn new things without forgetting old ones.
Online Class-Incremental Learning (OCIL) enables models to learn continuously from non-i.i.d. data streams and samples of the data streams can be seen only once, making it more suitable for real-world scenarios compared to offline learning. However, OCIL faces two key challenges: maintaining model stability under strict memory constraints and ensuring adaptability to new tasks. Under stricter memory constraints, current replay-based methods are less effective. While ensemble methods improve adaptability (plasticity), they often struggle with stability. To overcome these challenges, we propose a novel approach that enhances ensemble learning through a Global Workspace Model (GWM)-a shared, implicit memory that guides the learning of multiple student models. The GWM is formed by fusing the parameters of all students within each training batch, capturing the historical learning trajectory and serving as a dynamic anchor for knowledge consolidation. This fused model is then redistributed periodically to the students to stabilize learning and promote cross-task consistency. In addition, we introduce a multi-level collaborative distillation mechanism. This approach enforces peer-to-peer consistency among students and preserves historical knowledge by aligning each student with the GWM. As a result, student models remain adaptable to new tasks while maintaining previously learned knowledge, striking a better balance between stability and plasticity. Extensive experiments on three standard OCIL benchmarks show that our method delivers significant performance improvement for several OCIL models across various memory budgets.
Similar Papers
Inclusive Training Separation and Implicit Knowledge Interaction for Balanced Online Class-Incremental Learning
Machine Learning (CS)
Teaches computers to learn new things without forgetting old ones.
Large-Small Model Collaborative Framework for Federated Continual Learning
Machine Learning (CS)
Helps big AI learn new things without forgetting.
Being Strong Progressively! Enhancing Knowledge Distillation of Large Language Models through a Curriculum Learning Framework
Computation and Language
Makes big AI models smaller and faster.