Enhancing thermal infrared image colorization through reference-driven and contrastive learning approaches

Thermal infrared image colorization remains challenging due to limitations in existing methods, such as insufficient detail preservation and inaccurate color rendering. This paper presents a novel colorization approach that leverages reference images and contrastive learning to address these issues....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Infrared physics & technology 2025-03, Vol.145, p.105675, Article 105675
Hauptverfasser: Zhan, Weida, Shi, Mingkai, Chen, Yu, Zhang, Jingwen, Zhang, Cong, Han, Deng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Thermal infrared image colorization remains challenging due to limitations in existing methods, such as insufficient detail preservation and inaccurate color rendering. This paper presents a novel colorization approach that leverages reference images and contrastive learning to address these issues. Our model employs a dual-encoder generator architecture, allowing for detailed feature extraction from both infrared and reference images to enable precise color transfer. Key modules, including the Multi-Receptive Field Feature Integration Module (MFIM) and Channel–Spatial Feature Enhancement Module (CSFEM), enhance feature extraction and integration, while the Improved Stop-Gradient Attention Module (ISGA) ensures accurate feature alignment. A composite loss function combining adversarial, perceptual, and contrastive losses further refines the model’s output. Experimental results on benchmark datasets show that this method significantly improves colorization quality, generating visually realistic and detailed images, thus advancing applications in post-processing, object detection, and scene analysis within the infrared domain. •Proposed a dual-branch model combining contrastive and self-supervised learning.•Introduced ISGA to enhance feature fusion with improved stop-gradient attention.•Introduced MFIM for multi-scale feature extraction and enhanced perception.•Developed CSFEM to adaptively enhance channel and spatial feature representation.•Designed a composite loss to reduce detail loss, color distortion, and semantic blurring.
ISSN:1350-4495
DOI:10.1016/j.infrared.2024.105675