First-order State Space Model for Lightweight Image Super-resolution
By: Yujie Zhu , Xinyi Zhang , Yekai Lu and more
Potential Business Impact:
Makes pictures clearer with smarter computer math.
State space models (SSMs), particularly Mamba, have shown promise in NLP tasks and are increasingly applied to vision tasks. However, most Mamba-based vision models focus on network architecture and scan paths, with little attention to the SSM module. In order to explore the potential of SSMs, we modified the calculation process of SSM without increasing the number of parameters to improve the performance on lightweight super-resolution tasks. In this paper, we introduce the First-order State Space Model (FSSM) to improve the original Mamba module, enhancing performance by incorporating token correlations. We apply a first-order hold condition in SSMs, derive the new discretized form, and analyzed cumulative error. Extensive experimental results demonstrate that FSSM improves the performance of MambaIR on five benchmark datasets without additionally increasing the number of parameters, and surpasses current lightweight SR methods, achieving state-of-the-art results.
Similar Papers
X-VMamba: Explainable Vision Mamba
CV and Pattern Recognition
Shows how computer vision "sees" medical images.
SparseSSM: Efficient Selective Structured State Space Models Can Be Pruned in One-Shot
Machine Learning (CS)
Makes big AI models smaller without losing smarts.
SasMamba: A Lightweight Structure-Aware Stride State Space Model for 3D Human Pose Estimation
CV and Pattern Recognition
Helps computers understand human body movements better.