Scaling Spike-driven Transformer with Efficient Spike Firing Approximation Training
The ambition of brain-inspired Spiking Neural Networks (SNNs) is to become a low-power alternative to traditional Artificial Neural Networks (ANNs). This work addresses two major challenges in realizing this vision: the performance gap between SNNs and ANNs, and the high training costs of SNNs. We i...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The ambition of brain-inspired Spiking Neural Networks (SNNs) is to become a
low-power alternative to traditional Artificial Neural Networks (ANNs). This
work addresses two major challenges in realizing this vision: the performance
gap between SNNs and ANNs, and the high training costs of SNNs. We identify
intrinsic flaws in spiking neurons caused by binary firing mechanisms and
propose a Spike Firing Approximation (SFA) method using integer training and
spike-driven inference. This optimizes the spike firing pattern of spiking
neurons, enhancing efficient training, reducing power consumption, improving
performance, enabling easier scaling, and better utilizing neuromorphic chips.
We also develop an efficient spike-driven Transformer architecture and a
spike-masked autoencoder to prevent performance degradation during SNN scaling.
On ImageNet-1k, we achieve state-of-the-art top-1 accuracy of 78.5\%, 79.8\%,
84.0\%, and 86.2\% with models containing 10M, 19M, 83M, and 173M parameters,
respectively. For instance, the 10M model outperforms the best existing SNN by
7.2\% on ImageNet, with training time acceleration and inference energy
efficiency improved by 4.5$\times$ and 3.9$\times$, respectively. We validate
the effectiveness and efficiency of the proposed method across various tasks,
including object detection, semantic segmentation, and neuromorphic vision
tasks. This work enables SNNs to match ANN performance while maintaining the
low-power advantage, marking a significant step towards SNNs as a general
visual backbone. Code is available at
https://github.com/BICLab/Spike-Driven-Transformer-V3. |
---|---|
DOI: | 10.48550/arxiv.2411.16061 |