Deep Cross Spectral Stereo Matching Using Multi-Spectral Image Fusion
Cross spectral stereo matching aims to estimate disparity from color (RGB) and near-infrared (NIR) image pairs. The main difference from traditional stereo matching is that there is a big gap between two spectral bands, which makes cross spectral stereo matching a challenging task. In this letter, w...
Gespeichert in:
Veröffentlicht in: | IEEE robotics and automation letters 2022-04, Vol.7 (2), p.5373-5380 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Cross spectral stereo matching aims to estimate disparity from color (RGB) and near-infrared (NIR) image pairs. The main difference from traditional stereo matching is that there is a big gap between two spectral bands, which makes cross spectral stereo matching a challenging task. In this letter, we propose deep cross spectral stereo matching using multi-spectral image fusion. We adopt unsupervised learning to consider no ground truth in cross spectral stereo matching. We perform multi-spectral image fusion after cross spectral image translation to bridge the spectral gap between two images. First, we extract features from input RGB and NIR images to get fusion stereo pairs. Second, we get stereo correspondence robust to disparity variation based on parallax attention. In the loss function, we combine four losses: view reconstruction, material aware matching, cycle disparity consistency, and smoothness. We use a view reconstruction loss for spectral translation and fusion to warp stereo image pairs for appearance matching, while we use a material aware matching loss to take material property into consideration. Moreover, we utilize a cycle disparity consistency loss for disparity consistency between left and right predictions, and use a smoothness loss to enforce disparity smoothness. Experimental results show that the proposed network successfully estimates disparity with adaptability to materials and outperforms state-of-the-art models in terms of visual quality and quantitative measurements. |
---|---|
ISSN: | 2377-3766 2377-3766 |
DOI: | 10.1109/LRA.2022.3155202 |