Real-World Scene Image Enhancement with Contrastive Domain Adaptation Learning

Image enhancement methods leveraging learning-based approaches have demonstrated impressive results when trained on synthetic degraded-clear image pairs. However, when deployed in real-world scenarios, such models often suffer significant performance degradation due to the inherent domain gap betwee...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:ACM transactions on multimedia computing communications and applications 2024-12, Vol.20 (12), p.1-23, Article 371
Hauptverfasser: Zhang, Yongheng, Cai, Yuanqiang, Yan, Danfeng, Lin, Rongheng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Image enhancement methods leveraging learning-based approaches have demonstrated impressive results when trained on synthetic degraded-clear image pairs. However, when deployed in real-world scenarios, such models often suffer significant performance degradation due to the inherent domain gap between synthetic and real degradations. To bridge this gap, we propose a novel Two-stage Contrastive Domain Adaptation image Enhancement (TCDAE) framework consisting of two key strategies: (1) Synthetic-to-Real Domain Transfer Learning (S2R-DTL) that effectively translates images from the synthetic degraded domain to the real degraded domain, aligning the domains at the pixel level, and (2) Degraded-to-Clear Domain Transfer Learning (D2C-DTL) that further adapts the enhancement model from the synthetic to the real domain by translating images from the real degraded domain to the real clean domain in both supervised and unsupervised branches. A unique aspect of our approach is the integration of a Domain Noise Contrastive Estimation (DoNCE) loss in both learning strategies. This specialized loss formulation enables TCDAE to robustly translate images across domains, even in scenarios lacking strong positive examples. Consequently, our framework can generate enhanced images with natural, realistic appearances akin to real clear images. Comprehensive experiments on real-world degraded scenes across diverse tasks, including dehazing, deraining, and deblurring, demonstrate the superiority of TCDAE over state-of-the-art methods, achieving improved visual quality, quantitative metrics, and downstream task performance.
ISSN:1551-6857
1551-6865
DOI:10.1145/3694973