Segmentation based 6D pose estimation using integrated shape pattern and RGB information

Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene seg...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Pattern analysis and applications : PAA 2022-11, Vol.25 (4), p.1055-1073
Hauptverfasser: Gu, Chaochen, Feng, Qi, Lu, Changsheng, Zhao, Shuxin, Xu, Rui
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1073
container_issue 4
container_start_page 1055
container_title Pattern analysis and applications : PAA
container_volume 25
creator Gu, Chaochen
Feng, Qi
Lu, Changsheng
Zhao, Shuxin
Xu, Rui
description Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.
doi_str_mv 10.1007/s10044-022-01078-z
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2726451674</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2726451674</sourcerecordid><originalsourceid>FETCH-LOGICAL-c270t-1c8071ade73e8c8d819c1ae4c957c2a22c3dc10abfc8a0c382af0e40c3d33deb3</originalsourceid><addsrcrecordid>eNp9kE1LAzEQhoMoWKt_wFPAczRf26RHrVqFguAH9BbSZHbdYrNrkh7srze6ojcvMwPzvPPxInTK6DmjVF2kEqUklHNCGVWa7PbQiEkhiKqq5f5vLdkhOkppTakQgusRWj5Bs4GQbW67gFc2gceTa9x3CTCk3G6Gxja1ocFtyNBEmwuTXm0PuLc5QwzYBo8f51cFqLs4SI7RQW3fEpz85DF6ub15nt2RxcP8fna5II4rmglzmipmPSgB2mmv2dQxC9JNK-W45dwJ7xi1q9ppS53Q3NYUZKm8EB5WYozOhrl97N635WSz7rYxlJWGKz6RFZsoWSg-UC52KUWoTR_Lb_HDMGq-HDSDg6Y4aL4dNLsiEoMoFTg0EP9G_6P6BGIEdY4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2726451674</pqid></control><display><type>article</type><title>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</title><source>SpringerLink Journals - AutoHoldings</source><creator>Gu, Chaochen ; Feng, Qi ; Lu, Changsheng ; Zhao, Shuxin ; Xu, Rui</creator><creatorcontrib>Gu, Chaochen ; Feng, Qi ; Lu, Changsheng ; Zhao, Shuxin ; Xu, Rui</creatorcontrib><description>Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.</description><identifier>ISSN: 1433-7541</identifier><identifier>EISSN: 1433-755X</identifier><identifier>DOI: 10.1007/s10044-022-01078-z</identifier><language>eng</language><publisher>London: Springer London</publisher><subject>Color imagery ; Computer Science ; Feature extraction ; Image segmentation ; Industrial and Commercial Application ; Machine learning ; Object recognition ; Pattern Recognition ; Pose estimation ; Representations ; Three dimensional models</subject><ispartof>Pattern analysis and applications : PAA, 2022-11, Vol.25 (4), p.1055-1073</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c270t-1c8071ade73e8c8d819c1ae4c957c2a22c3dc10abfc8a0c382af0e40c3d33deb3</cites><orcidid>0000-0002-9748-7139</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10044-022-01078-z$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10044-022-01078-z$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Gu, Chaochen</creatorcontrib><creatorcontrib>Feng, Qi</creatorcontrib><creatorcontrib>Lu, Changsheng</creatorcontrib><creatorcontrib>Zhao, Shuxin</creatorcontrib><creatorcontrib>Xu, Rui</creatorcontrib><title>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</title><title>Pattern analysis and applications : PAA</title><addtitle>Pattern Anal Applic</addtitle><description>Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.</description><subject>Color imagery</subject><subject>Computer Science</subject><subject>Feature extraction</subject><subject>Image segmentation</subject><subject>Industrial and Commercial Application</subject><subject>Machine learning</subject><subject>Object recognition</subject><subject>Pattern Recognition</subject><subject>Pose estimation</subject><subject>Representations</subject><subject>Three dimensional models</subject><issn>1433-7541</issn><issn>1433-755X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kE1LAzEQhoMoWKt_wFPAczRf26RHrVqFguAH9BbSZHbdYrNrkh7srze6ojcvMwPzvPPxInTK6DmjVF2kEqUklHNCGVWa7PbQiEkhiKqq5f5vLdkhOkppTakQgusRWj5Bs4GQbW67gFc2gceTa9x3CTCk3G6Gxja1ocFtyNBEmwuTXm0PuLc5QwzYBo8f51cFqLs4SI7RQW3fEpz85DF6ub15nt2RxcP8fna5II4rmglzmipmPSgB2mmv2dQxC9JNK-W45dwJ7xi1q9ppS53Q3NYUZKm8EB5WYozOhrl97N635WSz7rYxlJWGKz6RFZsoWSg-UC52KUWoTR_Lb_HDMGq-HDSDg6Y4aL4dNLsiEoMoFTg0EP9G_6P6BGIEdY4</recordid><startdate>20221101</startdate><enddate>20221101</enddate><creator>Gu, Chaochen</creator><creator>Feng, Qi</creator><creator>Lu, Changsheng</creator><creator>Zhao, Shuxin</creator><creator>Xu, Rui</creator><general>Springer London</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-9748-7139</orcidid></search><sort><creationdate>20221101</creationdate><title>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</title><author>Gu, Chaochen ; Feng, Qi ; Lu, Changsheng ; Zhao, Shuxin ; Xu, Rui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c270t-1c8071ade73e8c8d819c1ae4c957c2a22c3dc10abfc8a0c382af0e40c3d33deb3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Color imagery</topic><topic>Computer Science</topic><topic>Feature extraction</topic><topic>Image segmentation</topic><topic>Industrial and Commercial Application</topic><topic>Machine learning</topic><topic>Object recognition</topic><topic>Pattern Recognition</topic><topic>Pose estimation</topic><topic>Representations</topic><topic>Three dimensional models</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Gu, Chaochen</creatorcontrib><creatorcontrib>Feng, Qi</creatorcontrib><creatorcontrib>Lu, Changsheng</creatorcontrib><creatorcontrib>Zhao, Shuxin</creatorcontrib><creatorcontrib>Xu, Rui</creatorcontrib><collection>CrossRef</collection><jtitle>Pattern analysis and applications : PAA</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gu, Chaochen</au><au>Feng, Qi</au><au>Lu, Changsheng</au><au>Zhao, Shuxin</au><au>Xu, Rui</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</atitle><jtitle>Pattern analysis and applications : PAA</jtitle><stitle>Pattern Anal Applic</stitle><date>2022-11-01</date><risdate>2022</risdate><volume>25</volume><issue>4</issue><spage>1055</spage><epage>1073</epage><pages>1055-1073</pages><issn>1433-7541</issn><eissn>1433-755X</eissn><abstract>Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.</abstract><cop>London</cop><pub>Springer London</pub><doi>10.1007/s10044-022-01078-z</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-9748-7139</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1433-7541
ispartof Pattern analysis and applications : PAA, 2022-11, Vol.25 (4), p.1055-1073
issn 1433-7541
1433-755X
language eng
recordid cdi_proquest_journals_2726451674
source SpringerLink Journals - AutoHoldings
subjects Color imagery
Computer Science
Feature extraction
Image segmentation
Industrial and Commercial Application
Machine learning
Object recognition
Pattern Recognition
Pose estimation
Representations
Three dimensional models
title Segmentation based 6D pose estimation using integrated shape pattern and RGB information
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T20%3A58%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Segmentation%20based%206D%20pose%20estimation%20using%20integrated%20shape%20pattern%20and%20RGB%20information&rft.jtitle=Pattern%20analysis%20and%20applications%20:%20PAA&rft.au=Gu,%20Chaochen&rft.date=2022-11-01&rft.volume=25&rft.issue=4&rft.spage=1055&rft.epage=1073&rft.pages=1055-1073&rft.issn=1433-7541&rft.eissn=1433-755X&rft_id=info:doi/10.1007/s10044-022-01078-z&rft_dat=%3Cproquest_cross%3E2726451674%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2726451674&rft_id=info:pmid/&rfr_iscdi=true