MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks
Point cloud data can reflect more information about the real 3D space, which has gained increasing attention in computer vision field. But the unstructured and unordered nature of point clouds poses many challenges in their study. How to learn the global features of the point cloud in the original p...
Gespeichert in:
Veröffentlicht in: | Machine vision and applications 2024-09, Vol.35 (5), p.111, Article 111 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 5 |
container_start_page | 111 |
container_title | Machine vision and applications |
container_volume | 35 |
creator | Liu, Wei Lu, Yisheng Zhang, Tao |
description | Point cloud data can reflect more information about the real 3D space, which has gained increasing attention in computer vision field. But the unstructured and unordered nature of point clouds poses many challenges in their study. How to learn the global features of the point cloud in the original point cloud is a problem that has been accompanied by the research. In the research based on the structure of the encoder and decoder, many researchers focus on designing the encoder to better extract features, and do not further explore more globally representative features according to the features of the encoder and decoder. To solve this problem, we propose the MFFAE-Net method, which aims to obtain more globally representative point cloud features by using the feature learning of encoder decoder stage.Our method first enhances the feature information of the input point cloud by merging the information of its neighboring points, which is helpful for the following point cloud feature extraction work. Secondly, the channel attention module is used to further process the extracted features, so as to highlight the role of important channels in the features. Finally, we fuse features of different scales from encoding features and decoding features as well as features of the same scale, so as to obtain more global point cloud features, which will help improve the segmentation results of point clouds. Experimental results show that the method performs well on some objects in S3DIS dataset and Toronto3d dataset. |
doi_str_mv | 10.1007/s00138-024-01589-1 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3092083319</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3092083319</sourcerecordid><originalsourceid>FETCH-LOGICAL-c200t-6af071796abde38523e3842ee8ce8e5c178e72c335c10da66e991e3b42f3f0513</originalsourceid><addsrcrecordid>eNp9kM1OwzAQhC0EEqXwApwscTas7fxyq6oWkApc4Gy5zqakJE6xHSHeHocgceOyO9bOfJaGkEsO1xwgv_EAXBYMRMKAp0XJ-BGZ8UQKxvOsPCYzKKMuoBSn5Mz7PQAkeZ7MSHhcrxcr9oThlnrstA2NiWLXoQ06NL2lfU0PfWMDNW0_VJ4OvrE72g1taJg3ukVaow6DizueYkDbiuoQImB8oX3T1uDIoxbDZ-_e_Tk5qXXr8eJ3z8nrevWyvGeb57uH5WLDjAAILNM15DwvM72tUBapkHEmArEwWGBqeF5gLoyUUUKlswzLkqPcJqKWNaRczsnVxD24_mNAH9S-H5yNXyoZq4BCSl5Gl5hcxvXeO6zVwTWddl-KgxrbVVO7KrarftpVI1pOIR_NdofuD_1P6hv7yn5l</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3092083319</pqid></control><display><type>article</type><title>MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks</title><source>SpringerLink Journals - AutoHoldings</source><creator>Liu, Wei ; Lu, Yisheng ; Zhang, Tao</creator><creatorcontrib>Liu, Wei ; Lu, Yisheng ; Zhang, Tao</creatorcontrib><description>Point cloud data can reflect more information about the real 3D space, which has gained increasing attention in computer vision field. But the unstructured and unordered nature of point clouds poses many challenges in their study. How to learn the global features of the point cloud in the original point cloud is a problem that has been accompanied by the research. In the research based on the structure of the encoder and decoder, many researchers focus on designing the encoder to better extract features, and do not further explore more globally representative features according to the features of the encoder and decoder. To solve this problem, we propose the MFFAE-Net method, which aims to obtain more globally representative point cloud features by using the feature learning of encoder decoder stage.Our method first enhances the feature information of the input point cloud by merging the information of its neighboring points, which is helpful for the following point cloud feature extraction work. Secondly, the channel attention module is used to further process the extracted features, so as to highlight the role of important channels in the features. Finally, we fuse features of different scales from encoding features and decoding features as well as features of the same scale, so as to obtain more global point cloud features, which will help improve the segmentation results of point clouds. Experimental results show that the method performs well on some objects in S3DIS dataset and Toronto3d dataset.</description><identifier>ISSN: 0932-8092</identifier><identifier>EISSN: 1432-1769</identifier><identifier>DOI: 10.1007/s00138-024-01589-1</identifier><language>eng</language><publisher>Berlin/Heidelberg: Springer Berlin Heidelberg</publisher><subject>Automation ; Cameras ; Coders ; Communications Engineering ; Computer Science ; Computer vision ; Datasets ; Decoding ; Feature extraction ; Image Processing and Computer Vision ; Image segmentation ; Methods ; Networks ; Pattern Recognition ; Semantic segmentation ; Semantics ; Sensors ; Three dimensional models ; Unstructured data</subject><ispartof>Machine vision and applications, 2024-09, Vol.35 (5), p.111, Article 111</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2024. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c200t-6af071796abde38523e3842ee8ce8e5c178e72c335c10da66e991e3b42f3f0513</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s00138-024-01589-1$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s00138-024-01589-1$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Liu, Wei</creatorcontrib><creatorcontrib>Lu, Yisheng</creatorcontrib><creatorcontrib>Zhang, Tao</creatorcontrib><title>MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks</title><title>Machine vision and applications</title><addtitle>Machine Vision and Applications</addtitle><description>Point cloud data can reflect more information about the real 3D space, which has gained increasing attention in computer vision field. But the unstructured and unordered nature of point clouds poses many challenges in their study. How to learn the global features of the point cloud in the original point cloud is a problem that has been accompanied by the research. In the research based on the structure of the encoder and decoder, many researchers focus on designing the encoder to better extract features, and do not further explore more globally representative features according to the features of the encoder and decoder. To solve this problem, we propose the MFFAE-Net method, which aims to obtain more globally representative point cloud features by using the feature learning of encoder decoder stage.Our method first enhances the feature information of the input point cloud by merging the information of its neighboring points, which is helpful for the following point cloud feature extraction work. Secondly, the channel attention module is used to further process the extracted features, so as to highlight the role of important channels in the features. Finally, we fuse features of different scales from encoding features and decoding features as well as features of the same scale, so as to obtain more global point cloud features, which will help improve the segmentation results of point clouds. Experimental results show that the method performs well on some objects in S3DIS dataset and Toronto3d dataset.</description><subject>Automation</subject><subject>Cameras</subject><subject>Coders</subject><subject>Communications Engineering</subject><subject>Computer Science</subject><subject>Computer vision</subject><subject>Datasets</subject><subject>Decoding</subject><subject>Feature extraction</subject><subject>Image Processing and Computer Vision</subject><subject>Image segmentation</subject><subject>Methods</subject><subject>Networks</subject><subject>Pattern Recognition</subject><subject>Semantic segmentation</subject><subject>Semantics</subject><subject>Sensors</subject><subject>Three dimensional models</subject><subject>Unstructured data</subject><issn>0932-8092</issn><issn>1432-1769</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kM1OwzAQhC0EEqXwApwscTas7fxyq6oWkApc4Gy5zqakJE6xHSHeHocgceOyO9bOfJaGkEsO1xwgv_EAXBYMRMKAp0XJ-BGZ8UQKxvOsPCYzKKMuoBSn5Mz7PQAkeZ7MSHhcrxcr9oThlnrstA2NiWLXoQ06NL2lfU0PfWMDNW0_VJ4OvrE72g1taJg3ukVaow6DizueYkDbiuoQImB8oX3T1uDIoxbDZ-_e_Tk5qXXr8eJ3z8nrevWyvGeb57uH5WLDjAAILNM15DwvM72tUBapkHEmArEwWGBqeF5gLoyUUUKlswzLkqPcJqKWNaRczsnVxD24_mNAH9S-H5yNXyoZq4BCSl5Gl5hcxvXeO6zVwTWddl-KgxrbVVO7KrarftpVI1pOIR_NdofuD_1P6hv7yn5l</recordid><startdate>20240901</startdate><enddate>20240901</enddate><creator>Liu, Wei</creator><creator>Lu, Yisheng</creator><creator>Zhang, Tao</creator><general>Springer Berlin Heidelberg</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20240901</creationdate><title>MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks</title><author>Liu, Wei ; Lu, Yisheng ; Zhang, Tao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c200t-6af071796abde38523e3842ee8ce8e5c178e72c335c10da66e991e3b42f3f0513</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Automation</topic><topic>Cameras</topic><topic>Coders</topic><topic>Communications Engineering</topic><topic>Computer Science</topic><topic>Computer vision</topic><topic>Datasets</topic><topic>Decoding</topic><topic>Feature extraction</topic><topic>Image Processing and Computer Vision</topic><topic>Image segmentation</topic><topic>Methods</topic><topic>Networks</topic><topic>Pattern Recognition</topic><topic>Semantic segmentation</topic><topic>Semantics</topic><topic>Sensors</topic><topic>Three dimensional models</topic><topic>Unstructured data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Liu, Wei</creatorcontrib><creatorcontrib>Lu, Yisheng</creatorcontrib><creatorcontrib>Zhang, Tao</creatorcontrib><collection>CrossRef</collection><jtitle>Machine vision and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Liu, Wei</au><au>Lu, Yisheng</au><au>Zhang, Tao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks</atitle><jtitle>Machine vision and applications</jtitle><stitle>Machine Vision and Applications</stitle><date>2024-09-01</date><risdate>2024</risdate><volume>35</volume><issue>5</issue><spage>111</spage><pages>111-</pages><artnum>111</artnum><issn>0932-8092</issn><eissn>1432-1769</eissn><abstract>Point cloud data can reflect more information about the real 3D space, which has gained increasing attention in computer vision field. But the unstructured and unordered nature of point clouds poses many challenges in their study. How to learn the global features of the point cloud in the original point cloud is a problem that has been accompanied by the research. In the research based on the structure of the encoder and decoder, many researchers focus on designing the encoder to better extract features, and do not further explore more globally representative features according to the features of the encoder and decoder. To solve this problem, we propose the MFFAE-Net method, which aims to obtain more globally representative point cloud features by using the feature learning of encoder decoder stage.Our method first enhances the feature information of the input point cloud by merging the information of its neighboring points, which is helpful for the following point cloud feature extraction work. Secondly, the channel attention module is used to further process the extracted features, so as to highlight the role of important channels in the features. Finally, we fuse features of different scales from encoding features and decoding features as well as features of the same scale, so as to obtain more global point cloud features, which will help improve the segmentation results of point clouds. Experimental results show that the method performs well on some objects in S3DIS dataset and Toronto3d dataset.</abstract><cop>Berlin/Heidelberg</cop><pub>Springer Berlin Heidelberg</pub><doi>10.1007/s00138-024-01589-1</doi></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0932-8092 |
ispartof | Machine vision and applications, 2024-09, Vol.35 (5), p.111, Article 111 |
issn | 0932-8092 1432-1769 |
language | eng |
recordid | cdi_proquest_journals_3092083319 |
source | SpringerLink Journals - AutoHoldings |
subjects | Automation Cameras Coders Communications Engineering Computer Science Computer vision Datasets Decoding Feature extraction Image Processing and Computer Vision Image segmentation Methods Networks Pattern Recognition Semantic segmentation Semantics Sensors Three dimensional models Unstructured data |
title | MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T19%3A02%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=MFFAE-Net:%20semantic%20segmentation%20of%20point%20clouds%20using%20multi-scale%20feature%20fusion%20and%20attention%20enhancement%20networks&rft.jtitle=Machine%20vision%20and%20applications&rft.au=Liu,%20Wei&rft.date=2024-09-01&rft.volume=35&rft.issue=5&rft.spage=111&rft.pages=111-&rft.artnum=111&rft.issn=0932-8092&rft.eissn=1432-1769&rft_id=info:doi/10.1007/s00138-024-01589-1&rft_dat=%3Cproquest_cross%3E3092083319%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3092083319&rft_id=info:pmid/&rfr_iscdi=true |