Interpretable Few-Shot Image Classification via Prototypical Concept-Guided Mixture of LoRA Experts
By: Zhong Ji , Rongshuai Wei , Jingren Liu and more
Potential Business Impact:
Helps computers understand pictures with less examples.
Self-Explainable Models (SEMs) rely on Prototypical Concept Learning (PCL) to enable their visual recognition processes more interpretable, but they often struggle in data-scarce settings where insufficient training samples lead to suboptimal performance.To address this limitation, we propose a Few-Shot Prototypical Concept Classification (FSPCC) framework that systematically mitigates two key challenges under low-data regimes: parametric imbalance and representation misalignment. Specifically, our approach leverages a Mixture of LoRA Experts (MoLE) for parameter-efficient adaptation, ensuring a balanced allocation of trainable parameters between the backbone and the PCL module.Meanwhile, cross-module concept guidance enforces tight alignment between the backbone's feature representations and the prototypical concept activation patterns.In addition, we incorporate a multi-level feature preservation strategy that fuses spatial and semantic cues across various layers, thereby enriching the learned representations and mitigating the challenges posed by limited data availability.Finally, to enhance interpretability and minimize concept overlap, we introduce a geometry-aware concept discrimination loss that enforces orthogonality among concepts, encouraging more disentangled and transparent decision boundaries.Experimental results on six popular benchmarks (CUB-200-2011, mini-ImageNet, CIFAR-FS, Stanford Cars, FGVC-Aircraft, and DTD) demonstrate that our approach consistently outperforms existing SEMs by a notable margin, with 4.2%-8.7% relative gains in 5-way 5-shot classification.These findings highlight the efficacy of coupling concept learning with few-shot adaptation to achieve both higher accuracy and clearer model interpretability, paving the way for more transparent visual recognition systems.
Similar Papers
MoTiC: Momentum Tightness and Contrast for Few-Shot Class-Incremental Learning
CV and Pattern Recognition
Teaches computers to learn new things without forgetting old ones.
Take a Peek: Efficient Encoder Adaptation for Few-Shot Semantic Segmentation via LoRA
CV and Pattern Recognition
Helps computers learn to identify new things quickly.
Towards Robust Visual Continual Learning with Multi-Prototype Supervision
CV and Pattern Recognition
Teaches computers to learn new things without forgetting.