Auto-Spikformer: Spikformer Architecture Search
The integration of self-attention mechanisms into Spiking Neural Networks (SNNs) has garnered considerable interest in the realm of advanced deep learning, primarily due to their biological properties. Recent advancements in SNN architecture, such as Spikformer, have demonstrated promising outcomes...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The integration of self-attention mechanisms into Spiking Neural Networks
(SNNs) has garnered considerable interest in the realm of advanced deep
learning, primarily due to their biological properties. Recent advancements in
SNN architecture, such as Spikformer, have demonstrated promising outcomes by
leveraging Spiking Self-Attention (SSA) and Spiking Patch Splitting (SPS)
modules. However, we observe that Spikformer may exhibit excessive energy
consumption, potentially attributable to redundant channels and blocks. To
mitigate this issue, we propose Auto-Spikformer, a one-shot Transformer
Architecture Search (TAS) method, which automates the quest for an optimized
Spikformer architecture. To facilitate the search process, we propose methods
Evolutionary SNN neurons (ESNN), which optimizes the SNN parameters, and apply
the previous method of weight entanglement supernet training, which optimizes
the Vision Transformer (ViT) parameters. Moreover, we propose an accuracy and
energy balanced fitness function $\mathcal{F}_{AEB}$ that jointly considers
both energy consumption and accuracy, and aims to find a Pareto optimal
combination that balances these two objectives. Our experimental results
demonstrate the effectiveness of Auto-Spikformer, which outperforms the
state-of-the-art method including CNN or ViT models that are manually or
automatically designed while significantly reducing energy consumption. |
---|---|
DOI: | 10.48550/arxiv.2306.00807 |