NLRNet: An Efficient Nonlocal Attention ResNet for Pansharpening

Remote sensing images often contain many similar components, such as buildings, roads, and water surfaces, which have similar spectra and spatial structures. Although convolutional neural networks (CNNs) based on residual learning can provide excellent performance in pansharpening, the existing meth...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on geoscience and remote sensing 2022, Vol.60, p.1-13
Hauptverfasser: Lei, Dajiang, Chen, Hao, Zhang, Liping, Li, Weisheng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Remote sensing images often contain many similar components, such as buildings, roads, and water surfaces, which have similar spectra and spatial structures. Although convolutional neural networks (CNNs) based on residual learning can provide excellent performance in pansharpening, the existing methods do not make full use of intrinsically similar information in images. Moreover, since the convolution operation is focused on the local region, even in a deep network, position-independent global information is difficult to obtain. In this article, an efficient nonlocal attention residual network (NLRNet) is proposed to capture the similar contextual dependencies of all pixels. Specifically, to reduce the difficulty of network training caused by the original nonlocal attention, we propose an efficient nonlocal attention (ENLA) mechanism and employ residual with zero initialization (ReZero) technology to make the signal easy to spread through the network. Furthermore, a spectral aggregation module (SpecAM) is proposed to generate fused images and adjust the corresponding spectral information. The experimental results for the QuickBird and WorldView3 data sets show that the proposed method is competitive with other advanced methods based on quality assessment and visual perception.
ISSN:0196-2892
1558-0644
DOI:10.1109/TGRS.2021.3067097