DAPnet: A Double Self-Attention Convolutional Network for Point Cloud Semantic Labeling

Airborne laser scanning (ALS) point clouds have complex structures, and their 3D semantic labeling has been a challenging task. It has three problems: 1) The difficulty of classifying point clouds around boundaries of objects from different classes, 2) the diversity of shapes within the same class,...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE journal of selected topics in applied earth observations and remote sensing 2021, Vol.14, p.9680-9691
Hauptverfasser: Chen, Li, Chen, Weiye, Xu, Zewei, Huang, Haozhe, Wang, Shaowen, Zhu, Qing, Li, Haifeng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Airborne laser scanning (ALS) point clouds have complex structures, and their 3D semantic labeling has been a challenging task. It has three problems: 1) The difficulty of classifying point clouds around boundaries of objects from different classes, 2) the diversity of shapes within the same class, and 3) the scale differences between classes. In this study, we propose a novel double self-attention convolutional network called the DAPnet. The double self-attention includes the point attention module (PAM) and the group attention module (GAM). For problem 1), the PAM can effectively assign different weights based on the relevance of point clouds in adjacent areas. Meanwhile, for problem 2), the GAM enhances the correlation between groups, i.e., grouped features within the same classes. To solve problem 3), we adopt a multiscale radius to construct the groups and concatenate extracted hierarchical features with the output of the corresponding upsampling process. Under the ISPRS 3D Semantic Labeling Contest dataset, the DAPnet outperforms the benchmark by 85.2% with an overall accuracy of 90.7%. By conducting ablation comparisons, we find that the PAM effectively improves the model than the GAM. The incorporation of the double self-attention module has an average of 7% improvement on the pre-class accuracy. Plus, the DAPnet consumes a similar training time to those without the attention modules for model convergence. The DAPnet can assign different weights to features based on the relevance between point clouds and their neighbors, which effectively improves classification performance.
ISSN:1939-1404
2151-1535
DOI:10.1109/JSTARS.2021.3113047