Mamba-X: An End-to-End Vision Mamba Accelerator for Edge Computing Devices
By: Dongho Yoon , Gungyu Lee , Jaewon Chang and more
Potential Business Impact:
Makes AI see better on small devices.
Transformers have proven effective in language modeling but are limited by high computational and memory demands that grow quadratically with input sequence length. State space models (SSMs) offer a promising alternative by reducing attention complexity from $O(L^2)$ to $O(L)$ while also lowering overall memory consumption. Vision Mamba adapts the SSM approach for computer vision tasks, achieving lower latency and memory consumption than traditional transformer models. However, deploying Vision Mamba on edge devices is challenging due to its sequential scan operations, which hinder GPU efficiency. We propose Mamba-X, an end-to-end Vision Mamba accelerator that includes a systolic scan array to maximize parallelism and minimize memory traffic, along with a hybrid, hardware-friendly quantization technique to reduce memory usage and improve hardware efficiency without sacrificing accuracy.
Similar Papers
eMamba: Efficient Acceleration Framework for Mamba Models in Edge Computing
Machine Learning (CS)
Makes smart devices run AI faster, using less power.
X-VMamba: Explainable Vision Mamba
CV and Pattern Recognition
Shows how computer vision "sees" medical images.
MMA: A Momentum Mamba Architecture for Human Activity Recognition with Inertial Sensors
Human-Computer Interaction
Helps computers understand body movements better.