Cross-Visual Attention Fusion Network with Dual-Constrained Marginal-Ranking for Visible-Infrared Person Re-Identification

Visible-Infrared Person re-identification(VI-REID) is extremely important for night-time surveillance applications. It is a challenging problem due to large cross-modality discrepancies and intra-modality variations caused by different illuminations, human poses, viewpoints, etc. In this paper, we p...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of physics. Conference series 2021-04, Vol.1880 (1), p.12033
Hauptverfasser: Su, Fang, Qi, Meibin, Chen, Cuiqun, Bo, Tan, Jiang, Jianguo
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Visible-Infrared Person re-identification(VI-REID) is extremely important for night-time surveillance applications. It is a challenging problem due to large cross-modality discrepancies and intra-modality variations caused by different illuminations, human poses, viewpoints, etc. In this paper, we propose a cross visual attention fusion dual-path neural network with dual-constrained marginal ranking(DCAF) to solve the problem. First, we utilize cross-visual attention to learn discriminative feature of high-level semantic information in their respective modals. Second, in order to establish the relationship between modals, we fuse attentional weight of two modals and add it into backpropagation to obtain those regions that are distinctive for classification. Third, a dual-constrained marginal-ranking loss is introduced to narrow the gap between different networks and to learn strongly the similarities between two modals. Extensive experiments demonstrate that the proposed approach effectively improves the performance of VI-REID task and remarkably outperforms the state-of-the-art methods.
ISSN:1742-6588
1742-6596
DOI:10.1088/1742-6596/1880/1/012033