AGMFusion: A Real-Time End-to-End Infrared and Visible Image Fusion Network Based on Adaptive Guidance Module

The goal of infrared and visible image fusion (IVIF) is to combine complementary information from the source images. However, due to the absence of ground truth, most fusion algorithms just utilize information from source images, which cannot provide targeted guidance for network learning, leading t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE sensors journal 2024-09, Vol.24 (17), p.28338-28350
Hauptverfasser: Liu, Shenghao, Lan, Xiaoxiong, Chen, Wenyong, Zhang, Zhiyong, Qiu, Changzhen
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The goal of infrared and visible image fusion (IVIF) is to combine complementary information from the source images. However, due to the absence of ground truth, most fusion algorithms just utilize information from source images, which cannot provide targeted guidance for network learning, leading to suboptimal outcomes. Moreover, numerous methods focus solely on modifying network architecture to improve fusion performance, without optimizing fusion algorithms from other perspectives. To tackle these problems, we propose a real-time end-to-end IVIF network based on an adaptive guidance module (AGMFusion). On the one hand, by combining the state-of-the-art (SOTA) methods and best fusion outcomes during the training process, the adaptive guidance module (AGM) effectively facilitates collaborative guidance for network training. Additionally, based on the AGM, we devise a loss function that includes content loss and guidance loss. We further balance these two loss components with the adaptive weight to boost the performance of our framework. On the other hand, AGMFusion is a lightweight image fusion framework that can generate highly perceptual fused images while maintaining excellent real-time performance, which can be potentially deployed as a preprocessing unit for various vision tasks. According to extensive comparative and generalization experiments, AGMFusion exceeds existing methods in visual appeal and quantitative metrics. Importantly, performance comparisons of numerous frameworks in running efficiency and object detection highlight the advantages of our approach. The source code will be released at https://github.com/liushh39/AGMFusion .
ISSN:1530-437X
1558-1748
DOI:10.1109/JSEN.2024.3426274