MFAGAN: A multiscale feature-attention generative adversarial network for infrared and visible image fusion
•A novel multiscale feature-attention generative adversarial network is proposed, which implements multi-path input based on multiscale image decomposition.•A learnable multiscale decomposition module is introduced to optimize feature extraction performance.•A novel feature-attention loss is introdu...
Gespeichert in:
Veröffentlicht in: | Infrared physics & technology 2023-09, Vol.133, p.104796, Article 104796 |
---|---|
Hauptverfasser: | , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | •A novel multiscale feature-attention generative adversarial network is proposed, which implements multi-path input based on multiscale image decomposition.•A learnable multiscale decomposition module is introduced to optimize feature extraction performance.•A novel feature-attention loss is introduced to guide the generator to pay more attention to the high-dimensional feature information of the images at different scales.•The proposed method has superiority in information balance, detail integrity, and visual understanding.
Infrared images contain salient target information and visible images contain texture information. The fusion of infrared and visible images makes images express better visual understanding. Further improvement in this area is obtained by the application of the Generative Adversarial Network (GAN). But some fusion results based on GAN will lose details from the source images. Besides, sometimes the fusion results are inclined to certain source images. To force the fusion results to retain more source features and balance the information from source images, a novel GAN called multiscale feature-attention generative adversarial network (MFAGAN) is proposed. First, the infrared and visible source images are decomposed into images at different scales. Then, the multiscale images are encoded and fused at the corresponding scale. Finally, the decoder generates fused images. The game between the generator and discriminator can make the information distribution of the fusion results more reasonable, but we further propose a new generator loss function called feature-attention loss. Feature-attention loss creates a criterion that measures the similarity of high-dimensional features between fused images and source images at each scale. Extensive experiments performed on two commonly used datasets show that MFAGAN obtains good results and has some superiority over other existing methods. |
---|---|
ISSN: | 1350-4495 1879-0275 |
DOI: | 10.1016/j.infrared.2023.104796 |