A Feature Fusion Network with Multiscale Adaptively Attentional for Object Detection in Complex Traffic Scenes

Currently, small object detection in complex traffic scenarios remains a major challenge due to variations in object shape, scales, and external environment, as well as a high interclass similarity. Although the latest algorithms achieve high accuracy, the results for small objects are still unsatis...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on intelligent vehicles 2024-10, p.1-14
Hauptverfasser: Cai, Fei, Qu, Zhong, Yin, Xuehui
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Currently, small object detection in complex traffic scenarios remains a major challenge due to variations in object shape, scales, and external environment, as well as a high interclass similarity. Although the latest algorithms achieve high accuracy, the results for small objects are still unsatisfactory. To obtain satisfactory small object detection results, the YOLOv8 algorithm is used to construct a feature fusion network with multiscale adaptively attentional (MAAFFNet). Firstly, the C2f module is modified to replace the simple summation operation with Multiscale Iterative Attentional Feature Fusion (MSIAFF), which enhances the ability of multi-scale aggregation of the C2f module and suppresses the interference of irrelevant information, resulting in a richer semantic feature map. Secondly, the adaptive spatial feature fusion (ASFF) module is introduced to deal with the feature maps of different scales, allocate the weights of spatial features of different levels, and enhance the consistency of semantic information between each layer of the feature maps, so as to improve the detail richness of the network. Finally, to achieve more accurate classification of small objects, a Slide loss function is introduced to achieve object classification and localization. Our proposed algorithm has been extensively experimented on KITTI, COCO-traffic and VOC 2012 datasets, respectively, comparing the proposed method to popular twostage and single-stage object detection algorithms, the proposed method performs well in terms of accuracy
ISSN:2379-8858
2379-8904
DOI:10.1109/TIV.2024.3476991