Toward understanding the effectiveness of attention mechanism
Attention mechanism (AM) is a widely used method for improving the performance of convolutional neural networks (CNNs) on computer vision tasks. Despite its pervasiveness, we have a poor understanding of what its effectiveness stems from. It is popularly believed that its effectiveness stems from th...
Gespeichert in:
Veröffentlicht in: | AIP advances 2023-03, Vol.13 (3), p.035019-035019-7 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Attention mechanism (AM) is a widely used method for improving the performance of convolutional neural networks (CNNs) on computer vision tasks. Despite its pervasiveness, we have a poor understanding of what its effectiveness stems from. It is popularly believed that its effectiveness stems from the visual attention explanation, i.e., attention weights indicate the importance of feature and AM advocates focusing on the important part of an input image rather than ingesting the entire input. However, we find only a weak consistency exists between the attention weights of features and their importance. We verify the feature map multiplication that brings about high-order non-linearity into CNNs is crucial for the effectiveness of AM. Furthermore, we show an essential impact of feature map multiplication on the learned surfaces of CNNs. With the high-order non-linearity, feature map multiplication plays a regularization role on CNNs, which makes the learned curves smoother and more stable in-between real samples (test/training samples in datasets). Thus, compared to vanilla CNNs, CNNs equipped with AM are more robust to noises and yield smaller model sensitivity scores, which is the reason for their better performance. |
---|---|
ISSN: | 2158-3226 2158-3226 |
DOI: | 10.1063/5.0141666 |