No-Extra Components Density Map Cropping Guided Object Detection in Aerial Images

Aerial images usually contain a large number of truncated and small objects, which poses a significant challenge for object detection. Existing methods have introduced additional learnable components in the pipeline and adopted multistage training approaches, but they have not solved the problem of...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on geoscience and remote sensing 2024, Vol.62, p.1-13
Hauptverfasser: Guo, Zhe, Bi, Guoling, Lv, Hengyi, Feng, Yang, Zhang, Yisa, Sun, Ming
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Aerial images usually contain a large number of truncated and small objects, which poses a significant challenge for object detection. Existing methods have introduced additional learnable components in the pipeline and adopted multistage training approaches, but they have not solved the problem of achieving end-to-end detection. To address this issue, we propose a novel no-extra components density map cropping (NE-CDMNet) method to utilize the spatial and contextual information between objects to improve detection performance. Furthermore, we introduce a new query selection (QS) scheme that utilizes confidence scores to select the top-K features from the encoder, helping the model better leverage the position information for extracting more comprehensive content features. Finally, we incorporate the local-global fusion (LGF) algorithm to combine the detection results from the original image and the density-cropped image. We conducted extensive experiments on two widely used public aerial datasets. Results reveal that the proposed method achieves the best performance compared with other state-of-the-art methods, whose 35.3% AP on the VisDrone-DET2019 dataset and 79.7% mAP on object detection in optical remote sensing image (DIOR) dataset, demonstrate the effectiveness of our method.
ISSN:0196-2892
1558-0644
DOI:10.1109/TGRS.2024.3481415