STAS: Spatio-Temporal Adaptive Computation Time for Spiking Transformers
By: Donghwa Kang , Doohyun Kim , Sang-Ki Ko and more
Potential Business Impact:
Makes AI see faster and use less power.
Spiking neural networks (SNNs) offer energy efficiency over artificial neural networks (ANNs) but suffer from high latency and computational overhead due to their multi-timestep operational nature. While various dynamic computation methods have been developed to mitigate this by targeting spatial, temporal, or architecture-specific redundancies, they remain fragmented. While the principles of adaptive computation time (ACT) offer a robust foundation for a unified approach, its application to SNN-based vision Transformers (ViTs) is hindered by two core issues: the violation of its temporal similarity prerequisite and a static architecture fundamentally unsuited for its principles. To address these challenges, we propose STAS (Spatio-Temporal Adaptive computation time for Spiking transformers), a framework that co-designs the static architecture and dynamic computation policy. STAS introduces an integrated spike patch splitting (I-SPS) module to establish temporal stability by creating a unified input representation, thereby solving the architectural problem of temporal dissimilarity. This stability, in turn, allows our adaptive spiking self-attention (A-SSA) module to perform two-dimensional token pruning across both spatial and temporal axes. Implemented on spiking Transformer architectures and validated on CIFAR-10, CIFAR-100, and ImageNet, STAS reduces energy consumption by up to 45.9%, 43.8%, and 30.1%, respectively, while simultaneously improving accuracy over SOTA models.
Similar Papers
STAA-SNN: Spatial-Temporal Attention Aggregator for Spiking Neural Networks
CV and Pattern Recognition
Helps computers learn faster and better from moving pictures.
SparseST: Exploiting Data Sparsity in Spatiotemporal Modeling and Prediction
Machine Learning (CS)
Makes smart AI work on small, cheap devices.
Timestep-Compressed Attack on Spiking Neural Networks through Timestep-Level Backpropagation
CV and Pattern Recognition
Makes AI attacks on brain-like computers faster.