Lightweight super-resolution via multi-group window self-attention and residual blueprint separable convolution
Benefiting from self-attention mechanism and convolutional operation, numerous lightweight Transformer-based single image super-resolution (SISR) approaches have made considerable breakthroughs in performance. Nevertheless, higher computational costs of self-attention and parametric overheads of con...
Gespeichert in:
Veröffentlicht in: | Multimedia systems 2024-10, Vol.30 (5), Article 276 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Benefiting from self-attention mechanism and convolutional operation, numerous lightweight Transformer-based single image super-resolution (SISR) approaches have made considerable breakthroughs in performance. Nevertheless, higher computational costs of self-attention and parametric overheads of convolution still limit the deployment of these methods on low-budget devices. To tackle the above issues, based on the sequentially cascaded attention and convolution-cooperative blocks (ACCBs), we propose an attention and convolution-cooperative network (ACCNet) for lightweight image super-resolution (SR). Specifically, in each ACCB, to alleviate the computational burden of self-attention mechanism, we propose a multi-group window self-attention (MGWSA), which achieves self-attention calculation on different groups of features via different window sizes. To implement a lightweight convolutional operation that assists self-attention for local feature extraction, we propose a residual blueprint separable convolution (RBSC), which combines the advantages of efficient convolution and residual learning. Additionally, an enhanced multi-layer perceptron (EMLP) is designed to strengthen the representation of features spatially. The contrast-aware channel attention (CCA) is introduced to exploit the beneficial interdependency among channels. Compared to other lightweight state-of-the-art SR methods, the better trade-off between model complexity and performance in our ACCNet is demonstrated by extensive experiments. |
---|---|
ISSN: | 0942-4962 1432-1882 |
DOI: | 10.1007/s00530-024-01501-x |