Class-Independent Increment: An Efficient Approach for Multi-label Class-Incremental Learning
By: Songlin Dong , Yuhang He , Zhengdong Zhou and more
Potential Business Impact:
Teaches computers to learn new things without forgetting old ones.
Current research on class-incremental learning primarily focuses on single-label classification tasks. However, real-world applications often involve multi-label scenarios, such as image retrieval and medical imaging. Therefore, this paper focuses on the challenging yet practical multi-label class-incremental learning (MLCIL) problem. In addition to the challenge of catastrophic forgetting, MLCIL encounters issues related to feature confusion, encompassing inter-session and intra-feature confusion. To address these problems, we propose a novel MLCIL approach called class-independent increment (CLIN). Specifically, in contrast to existing methods that extract image-level features, we propose a class-independent incremental network (CINet) to extract multiple class-level embeddings for multi-label samples. It learns and preserves the knowledge of different classes by constructing class-specific tokens. On this basis, we develop two novel loss functions, optimizing the learning of class-specific tokens and class-level embeddings, respectively. These losses aim to distinguish between new and old classes, further alleviating the problem of feature confusion. Extensive experiments on MS-COCO and PASCAL VOC datasets demonstrate the effectiveness of our method for improving recognition performance and mitigating forgetting on various MLCIL tasks.
Similar Papers
Specifying What You Know or Not for Multi-Label Class-Incremental Learning
Machine Learning (CS)
Teaches computers to learn new things without forgetting.
Unsupervised Incremental Learning Using Confidence-Based Pseudo-Labels
CV and Pattern Recognition
Teaches computers new things without labels.
Annotation-Free Class-Incremental Learning
Machine Learning (CS)
Teaches computers new things without labels.