Scalpel-SAM: A Semi-Supervised Paradigm for Adapting SAM to Infrared Small Object Detection
By: Zihan Liu , Xiangning Ren , Dezhang Kong and more
Potential Business Impact:
Teaches computers to find tiny things with less help.
Infrared small object detection urgently requires semi-supervised paradigms due to the high cost of annotation. However, existing methods like SAM face significant challenges of domain gaps, inability of encoding physical priors, and inherent architectural complexity. To address this, we designed a Hierarchical MoE Adapter consisting of four white-box neural operators. Building upon this core component, we propose a two-stage paradigm for knowledge distillation and transfer: (1) Prior-Guided Knowledge Distillation, where we use our MoE adapter and 10% of available fully supervised data to distill SAM into an expert teacher (Scalpel-SAM); and (2) Deployment-Oriented Knowledge Transfer, where we use Scalpel-SAM to generate pseudo labels for training lightweight and efficient downstream models. Experiments demonstrate that with minimal annotations, our paradigm enables downstream models to achieve performance comparable to, or even surpassing, their fully supervised counterparts. To our knowledge, this is the first semi-supervised paradigm that systematically addresses the data scarcity issue in IR-SOT using SAM as the teacher model.
Similar Papers
ST-SAM: SAM-Driven Self-Training Framework for Semi-Supervised Camouflaged Object Detection
CV and Pattern Recognition
Teaches computers to find hidden things with less help.
S^4M: Boosting Semi-Supervised Instance Segmentation with SAM
CV and Pattern Recognition
Teaches computers to find and label things in pictures.
DOD-SA: Infrared-Visible Decoupled Object Detection with Single-Modality Annotations
CV and Pattern Recognition
Helps cameras see in dark and light.