Learn Faster and Remember More: Balancing Exploration and Exploitation for Continual Test-time Adaptation
By: Pinci Yang , Peisong Wen , Ke Ma and more
Potential Business Impact:
Helps AI learn new things without forgetting old ones.
Plain English Summary
Imagine you have a smart camera that can recognize different types of dogs. This new method helps the camera learn to recognize new dog breeds it's never seen before, without forgetting the ones it already knows. This means the camera can get smarter over time and be more accurate in different situations, like recognizing a new breed of dog at a park you visit for the first time.
Continual Test-Time Adaptation (CTTA) aims to adapt a source pre-trained model to continually changing target domains during inference. As a fundamental principle, an ideal CTTA method should rapidly adapt to new domains (exploration) while retaining and exploiting knowledge from previously encountered domains to handle similar domains in the future. Despite significant advances, balancing exploration and exploitation in CTTA is still challenging: 1) Existing methods focus on adjusting predictions based on deep-layer outputs of neural networks. However, domain shifts typically affect shallow features, which are inefficient to be adjusted from deep predictions, leading to dilatory exploration; 2) A single model inevitably forgets knowledge of previous domains during the exploration, making it incapable of exploiting historical knowledge to handle similar future domains. To address these challenges, this paper proposes a mean teacher framework that strikes an appropriate Balance between Exploration and Exploitation (BEE) during the CTTA process. For the former challenge, we introduce a Multi-level Consistency Regularization (MCR) loss that aligns the intermediate features of the student and teacher models, accelerating adaptation to the current domain. For the latter challenge, we employ a Complementary Anchor Replay (CAR) mechanism to reuse historical checkpoints (anchors), recovering complementary knowledge for diverse domains. Experiments show that our method significantly outperforms state-of-the-art methods on several benchmarks, demonstrating its effectiveness for CTTA tasks.
Similar Papers
CTTA-T: Continual Test-Time Adaptation for Text Understanding via Teacher-Student with a Domain-aware and Generalized Teacher
Computation and Language
Helps computers understand new text as it changes.
Instance-Aware Test-Time Segmentation for Continual Domain Shifts
CV and Pattern Recognition
Helps AI see better as things change.
SloMo-Fast: Slow-Momentum and Fast-Adaptive Teachers for Source-Free Continual Test-Time Adaptation
Machine Learning (CS)
Keeps AI smart on new and old tasks.