Deep Cross Spectral Stereo Matching Using Multi-Spectral Image Fusion

Cross spectral stereo matching aims to estimate disparity from color (RGB) and near-infrared (NIR) image pairs. The main difference from traditional stereo matching is that there is a big gap between two spectral bands, which makes cross spectral stereo matching a challenging task. In this letter, w...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE robotics and automation letters 2022-04, Vol.7 (2), p.5373-5380
Hauptverfasser: Liang, Xiaolong, Jung, Cheolkon
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Cross spectral stereo matching aims to estimate disparity from color (RGB) and near-infrared (NIR) image pairs. The main difference from traditional stereo matching is that there is a big gap between two spectral bands, which makes cross spectral stereo matching a challenging task. In this letter, we propose deep cross spectral stereo matching using multi-spectral image fusion. We adopt unsupervised learning to consider no ground truth in cross spectral stereo matching. We perform multi-spectral image fusion after cross spectral image translation to bridge the spectral gap between two images. First, we extract features from input RGB and NIR images to get fusion stereo pairs. Second, we get stereo correspondence robust to disparity variation based on parallax attention. In the loss function, we combine four losses: view reconstruction, material aware matching, cycle disparity consistency, and smoothness. We use a view reconstruction loss for spectral translation and fusion to warp stereo image pairs for appearance matching, while we use a material aware matching loss to take material property into consideration. Moreover, we utilize a cycle disparity consistency loss for disparity consistency between left and right predictions, and use a smoothness loss to enforce disparity smoothness. Experimental results show that the proposed network successfully estimates disparity with adaptability to materials and outperforms state-of-the-art models in terms of visual quality and quantitative measurements.
ISSN:2377-3766
2377-3766
DOI:10.1109/LRA.2022.3155202