Segmentation based 6D pose estimation using integrated shape pattern and RGB information
Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene seg...
Gespeichert in:
Veröffentlicht in: | Pattern analysis and applications : PAA 2022-11, Vol.25 (4), p.1055-1073 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1073 |
---|---|
container_issue | 4 |
container_start_page | 1055 |
container_title | Pattern analysis and applications : PAA |
container_volume | 25 |
creator | Gu, Chaochen Feng, Qi Lu, Changsheng Zhao, Shuxin Xu, Rui |
description | Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation. |
doi_str_mv | 10.1007/s10044-022-01078-z |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2726451674</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2726451674</sourcerecordid><originalsourceid>FETCH-LOGICAL-c270t-1c8071ade73e8c8d819c1ae4c957c2a22c3dc10abfc8a0c382af0e40c3d33deb3</originalsourceid><addsrcrecordid>eNp9kE1LAzEQhoMoWKt_wFPAczRf26RHrVqFguAH9BbSZHbdYrNrkh7srze6ojcvMwPzvPPxInTK6DmjVF2kEqUklHNCGVWa7PbQiEkhiKqq5f5vLdkhOkppTakQgusRWj5Bs4GQbW67gFc2gceTa9x3CTCk3G6Gxja1ocFtyNBEmwuTXm0PuLc5QwzYBo8f51cFqLs4SI7RQW3fEpz85DF6ub15nt2RxcP8fna5II4rmglzmipmPSgB2mmv2dQxC9JNK-W45dwJ7xi1q9ppS53Q3NYUZKm8EB5WYozOhrl97N635WSz7rYxlJWGKz6RFZsoWSg-UC52KUWoTR_Lb_HDMGq-HDSDg6Y4aL4dNLsiEoMoFTg0EP9G_6P6BGIEdY4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2726451674</pqid></control><display><type>article</type><title>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</title><source>SpringerLink Journals - AutoHoldings</source><creator>Gu, Chaochen ; Feng, Qi ; Lu, Changsheng ; Zhao, Shuxin ; Xu, Rui</creator><creatorcontrib>Gu, Chaochen ; Feng, Qi ; Lu, Changsheng ; Zhao, Shuxin ; Xu, Rui</creatorcontrib><description>Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.</description><identifier>ISSN: 1433-7541</identifier><identifier>EISSN: 1433-755X</identifier><identifier>DOI: 10.1007/s10044-022-01078-z</identifier><language>eng</language><publisher>London: Springer London</publisher><subject>Color imagery ; Computer Science ; Feature extraction ; Image segmentation ; Industrial and Commercial Application ; Machine learning ; Object recognition ; Pattern Recognition ; Pose estimation ; Representations ; Three dimensional models</subject><ispartof>Pattern analysis and applications : PAA, 2022-11, Vol.25 (4), p.1055-1073</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c270t-1c8071ade73e8c8d819c1ae4c957c2a22c3dc10abfc8a0c382af0e40c3d33deb3</cites><orcidid>0000-0002-9748-7139</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10044-022-01078-z$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10044-022-01078-z$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Gu, Chaochen</creatorcontrib><creatorcontrib>Feng, Qi</creatorcontrib><creatorcontrib>Lu, Changsheng</creatorcontrib><creatorcontrib>Zhao, Shuxin</creatorcontrib><creatorcontrib>Xu, Rui</creatorcontrib><title>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</title><title>Pattern analysis and applications : PAA</title><addtitle>Pattern Anal Applic</addtitle><description>Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.</description><subject>Color imagery</subject><subject>Computer Science</subject><subject>Feature extraction</subject><subject>Image segmentation</subject><subject>Industrial and Commercial Application</subject><subject>Machine learning</subject><subject>Object recognition</subject><subject>Pattern Recognition</subject><subject>Pose estimation</subject><subject>Representations</subject><subject>Three dimensional models</subject><issn>1433-7541</issn><issn>1433-755X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kE1LAzEQhoMoWKt_wFPAczRf26RHrVqFguAH9BbSZHbdYrNrkh7srze6ojcvMwPzvPPxInTK6DmjVF2kEqUklHNCGVWa7PbQiEkhiKqq5f5vLdkhOkppTakQgusRWj5Bs4GQbW67gFc2gceTa9x3CTCk3G6Gxja1ocFtyNBEmwuTXm0PuLc5QwzYBo8f51cFqLs4SI7RQW3fEpz85DF6ub15nt2RxcP8fna5II4rmglzmipmPSgB2mmv2dQxC9JNK-W45dwJ7xi1q9ppS53Q3NYUZKm8EB5WYozOhrl97N635WSz7rYxlJWGKz6RFZsoWSg-UC52KUWoTR_Lb_HDMGq-HDSDg6Y4aL4dNLsiEoMoFTg0EP9G_6P6BGIEdY4</recordid><startdate>20221101</startdate><enddate>20221101</enddate><creator>Gu, Chaochen</creator><creator>Feng, Qi</creator><creator>Lu, Changsheng</creator><creator>Zhao, Shuxin</creator><creator>Xu, Rui</creator><general>Springer London</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-9748-7139</orcidid></search><sort><creationdate>20221101</creationdate><title>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</title><author>Gu, Chaochen ; Feng, Qi ; Lu, Changsheng ; Zhao, Shuxin ; Xu, Rui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c270t-1c8071ade73e8c8d819c1ae4c957c2a22c3dc10abfc8a0c382af0e40c3d33deb3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Color imagery</topic><topic>Computer Science</topic><topic>Feature extraction</topic><topic>Image segmentation</topic><topic>Industrial and Commercial Application</topic><topic>Machine learning</topic><topic>Object recognition</topic><topic>Pattern Recognition</topic><topic>Pose estimation</topic><topic>Representations</topic><topic>Three dimensional models</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Gu, Chaochen</creatorcontrib><creatorcontrib>Feng, Qi</creatorcontrib><creatorcontrib>Lu, Changsheng</creatorcontrib><creatorcontrib>Zhao, Shuxin</creatorcontrib><creatorcontrib>Xu, Rui</creatorcontrib><collection>CrossRef</collection><jtitle>Pattern analysis and applications : PAA</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gu, Chaochen</au><au>Feng, Qi</au><au>Lu, Changsheng</au><au>Zhao, Shuxin</au><au>Xu, Rui</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Segmentation based 6D pose estimation using integrated shape pattern and RGB information</atitle><jtitle>Pattern analysis and applications : PAA</jtitle><stitle>Pattern Anal Applic</stitle><date>2022-11-01</date><risdate>2022</risdate><volume>25</volume><issue>4</issue><spage>1055</spage><epage>1073</epage><pages>1055-1073</pages><issn>1433-7541</issn><eissn>1433-755X</eissn><abstract>Point cloud is currently the most typical representation in describing the 3D world. However, recognizing objects as well as the poses from point clouds is still a great challenge due to the property of disordered 3D data arrangement. In this paper, a unified deep learning framework for 3D scene segmentation and 6D object pose estimation is proposed. In order to accurately segment foreground objects, a novel shape pattern aggregation module called PointDoN is proposed, which could learn meaningful deep geometric representations from both Difference of Normals (DoN) and the initial spatial coordinates of point cloud. Our PointDoN is flexible to be applied to any convolutional networks and shows improvements in the popular tasks of point cloud classification and semantic segmentation. Once the objects are segmented, the range of point clouds for each object in the entire scene could be specified, which enables us to further estimate the 6D pose for each object within local region of interest. To acquire good estimate, we propose a new 6D pose estimation approach that incorporates both 2D and 3D features generated from RGB images and point clouds, respectively. Specifically, 3D features are extracted via a CNN-based architecture where the input is XYZ map converted from the initial point cloud. Experiments showed that our method could achieve satisfactory results on the publicly available point cloud datasets in both tasks of segmentation and 6D pose estimation.</abstract><cop>London</cop><pub>Springer London</pub><doi>10.1007/s10044-022-01078-z</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-9748-7139</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1433-7541 |
ispartof | Pattern analysis and applications : PAA, 2022-11, Vol.25 (4), p.1055-1073 |
issn | 1433-7541 1433-755X |
language | eng |
recordid | cdi_proquest_journals_2726451674 |
source | SpringerLink Journals - AutoHoldings |
subjects | Color imagery Computer Science Feature extraction Image segmentation Industrial and Commercial Application Machine learning Object recognition Pattern Recognition Pose estimation Representations Three dimensional models |
title | Segmentation based 6D pose estimation using integrated shape pattern and RGB information |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T20%3A58%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Segmentation%20based%206D%20pose%20estimation%20using%20integrated%20shape%20pattern%20and%20RGB%20information&rft.jtitle=Pattern%20analysis%20and%20applications%20:%20PAA&rft.au=Gu,%20Chaochen&rft.date=2022-11-01&rft.volume=25&rft.issue=4&rft.spage=1055&rft.epage=1073&rft.pages=1055-1073&rft.issn=1433-7541&rft.eissn=1433-755X&rft_id=info:doi/10.1007/s10044-022-01078-z&rft_dat=%3Cproquest_cross%3E2726451674%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2726451674&rft_id=info:pmid/&rfr_iscdi=true |