RA-MMIR : Multi-modal image registration by Robust Adaptive Variation Attention Gauge Field
Multi-modal image registration finds extensive application in high-level vision tasks. Especially in adverse conditions, multi-modal image registration and fusion is powerful for high-level visual analysis, high-level visual tasks commonly prioritize the analysis of crucial target regions within ima...
Gespeichert in:
Veröffentlicht in: | Information fusion 2024-05, Vol.105, p.102215, Article 102215 |
---|---|
Hauptverfasser: | , , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Multi-modal image registration finds extensive application in high-level vision tasks. Especially in adverse conditions, multi-modal image registration and fusion is powerful for high-level visual analysis, high-level visual tasks commonly prioritize the analysis of crucial target regions within images, however precise multi-modal image registration remains a challenge. To address this issue, we rethought the collaboration between image registration and high-level visual tasks, and propose a Robust Adaptive Variation Attention Gauge Field registration framework that allows for flexible attention to both the target regions and global areas. Among them, in order to improve the robustness and timeliness of the algorithm in extracting and describing target regions or global areas features in adverse conditions, we propose a Robust Adaptive Variation Attention for building the gauge field, and in order to make robust adaptive parameters to reach the global optimum, we propose a Quasi-Simulated Annealing method based on mini-batch. To make the spatial transformation better fit feature matching and image registration, we design a deep learning and modeling approach based on spatial similarity metrics. In high-level visual tasks-based experiments such as image fusion, object detection, and 3D reconstruction, our method had the best collaborative performance and showed the best registration results under adverse conditions. The code is at https://github.com/JuiHuiQ/RA-MMIR.
•We propose a Weak Boundary Constraints-based Flexible Matching strategy.•We propose a RAMM Attention to perform feature extraction and description.•We propose a Unified-VEM to minimize computational complexity.•We propose a SP-SSIM making transformation more consistent with real-world scene. |
---|---|
ISSN: | 1566-2535 1872-6305 |
DOI: | 10.1016/j.inffus.2023.102215 |