Revisiting the Effectiveness of 3D Object Recognition Benchmarks
Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding....
Gespeichert in:
Veröffentlicht in: | IEEE access 2023, Vol.11, p.111566-111580 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 111580 |
---|---|
container_issue | |
container_start_page | 111566 |
container_title | IEEE access |
container_volume | 11 |
creator | Song, Hyunsoo Lee, Seungkyu |
description | Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public. |
doi_str_mv | 10.1109/ACCESS.2023.3322433 |
format | Article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_10273253</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10273253</ieee_id><doaj_id>oai_doaj_org_article_254cf39418b14c3484a7a4ea4d40b499</doaj_id><sourcerecordid>2878510088</sourcerecordid><originalsourceid>FETCH-LOGICAL-c359t-649bf7e4ce8f14e61b22f18794e3e99a47f8adf4e145816a000a109ec982d9a43</originalsourceid><addsrcrecordid>eNpNUNFOwjAUXYwmEuQL9GGJz-Dae8faNxFRSUhIQJ-brruFIa64DhL_3uKIoS9tTs8599wTRbcsGTCWyIfReDxZLgc84TAA4BwBLqIOZ0PZhxSGl2fv66jn_SYJRwQozTrR44IOpS-bslrFzZriibVkmvJAFXkfOxvDczzPNwGLF2TcqgpUV8VPVJn1l64__U10ZfXWU-90d6OPl8n7-K0_m79Ox6NZ30Aqm_4QZW4zQkPCMqQhyzm3TGQSCUhKjZkVurBIDNOQTYeIOixHRgpehG_oRtPWt3B6o3Z1Gab_KKdL9Qe4eqV03ZRmS4qnaCxIZCJnaAAF6kwjaSwwyVHK4HXfeu1q970n36iN29dViK-4yETKQj8isKBlmdp5X5P9n8oSdWxetc2rY_Pq1HxQ3bWqkojOFDwDngL8Ap99fT8</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2878510088</pqid></control><display><type>article</type><title>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Song, Hyunsoo ; Lee, Seungkyu</creator><creatorcontrib>Song, Hyunsoo ; Lee, Seungkyu</creatorcontrib><description>Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2023.3322433</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Benchmark ; Benchmark testing ; Benchmarks ; Computer vision ; Data acquisition ; Data augmentation ; dataset ; Datasets ; Effectiveness ; Industrial applications ; Motion perception ; Object recognition ; point cloud ; Point cloud compression ; Scene analysis ; Solid modeling ; Task analysis ; Three dimensional bodies ; Three dimensional models ; Three dimensional motion ; Three-dimensional displays</subject><ispartof>IEEE access, 2023, Vol.11, p.111566-111580</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c359t-649bf7e4ce8f14e61b22f18794e3e99a47f8adf4e145816a000a109ec982d9a43</cites><orcidid>0009-0001-5279-0505 ; 0000-0002-9721-4093</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10273253$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2102,4024,27633,27923,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Song, Hyunsoo</creatorcontrib><creatorcontrib>Lee, Seungkyu</creatorcontrib><title>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</title><title>IEEE access</title><addtitle>Access</addtitle><description>Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.</description><subject>Benchmark</subject><subject>Benchmark testing</subject><subject>Benchmarks</subject><subject>Computer vision</subject><subject>Data acquisition</subject><subject>Data augmentation</subject><subject>dataset</subject><subject>Datasets</subject><subject>Effectiveness</subject><subject>Industrial applications</subject><subject>Motion perception</subject><subject>Object recognition</subject><subject>point cloud</subject><subject>Point cloud compression</subject><subject>Scene analysis</subject><subject>Solid modeling</subject><subject>Task analysis</subject><subject>Three dimensional bodies</subject><subject>Three dimensional models</subject><subject>Three dimensional motion</subject><subject>Three-dimensional displays</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUNFOwjAUXYwmEuQL9GGJz-Dae8faNxFRSUhIQJ-brruFIa64DhL_3uKIoS9tTs8599wTRbcsGTCWyIfReDxZLgc84TAA4BwBLqIOZ0PZhxSGl2fv66jn_SYJRwQozTrR44IOpS-bslrFzZriibVkmvJAFXkfOxvDczzPNwGLF2TcqgpUV8VPVJn1l64__U10ZfXWU-90d6OPl8n7-K0_m79Ox6NZ30Aqm_4QZW4zQkPCMqQhyzm3TGQSCUhKjZkVurBIDNOQTYeIOixHRgpehG_oRtPWt3B6o3Z1Gab_KKdL9Qe4eqV03ZRmS4qnaCxIZCJnaAAF6kwjaSwwyVHK4HXfeu1q970n36iN29dViK-4yETKQj8isKBlmdp5X5P9n8oSdWxetc2rY_Pq1HxQ3bWqkojOFDwDngL8Ap99fT8</recordid><startdate>2023</startdate><enddate>2023</enddate><creator>Song, Hyunsoo</creator><creator>Lee, Seungkyu</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0009-0001-5279-0505</orcidid><orcidid>https://orcid.org/0000-0002-9721-4093</orcidid></search><sort><creationdate>2023</creationdate><title>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</title><author>Song, Hyunsoo ; Lee, Seungkyu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c359t-649bf7e4ce8f14e61b22f18794e3e99a47f8adf4e145816a000a109ec982d9a43</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Benchmark</topic><topic>Benchmark testing</topic><topic>Benchmarks</topic><topic>Computer vision</topic><topic>Data acquisition</topic><topic>Data augmentation</topic><topic>dataset</topic><topic>Datasets</topic><topic>Effectiveness</topic><topic>Industrial applications</topic><topic>Motion perception</topic><topic>Object recognition</topic><topic>point cloud</topic><topic>Point cloud compression</topic><topic>Scene analysis</topic><topic>Solid modeling</topic><topic>Task analysis</topic><topic>Three dimensional bodies</topic><topic>Three dimensional models</topic><topic>Three dimensional motion</topic><topic>Three-dimensional displays</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Song, Hyunsoo</creatorcontrib><creatorcontrib>Lee, Seungkyu</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Song, Hyunsoo</au><au>Lee, Seungkyu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2023</date><risdate>2023</risdate><volume>11</volume><spage>111566</spage><epage>111580</epage><pages>111566-111580</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2023.3322433</doi><tpages>15</tpages><orcidid>https://orcid.org/0009-0001-5279-0505</orcidid><orcidid>https://orcid.org/0000-0002-9721-4093</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2023, Vol.11, p.111566-111580 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_ieee_primary_10273253 |
source | IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals |
subjects | Benchmark Benchmark testing Benchmarks Computer vision Data acquisition Data augmentation dataset Datasets Effectiveness Industrial applications Motion perception Object recognition point cloud Point cloud compression Scene analysis Solid modeling Task analysis Three dimensional bodies Three dimensional models Three dimensional motion Three-dimensional displays |
title | Revisiting the Effectiveness of 3D Object Recognition Benchmarks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T18%3A36%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Revisiting%20the%20Effectiveness%20of%203D%20Object%20Recognition%20Benchmarks&rft.jtitle=IEEE%20access&rft.au=Song,%20Hyunsoo&rft.date=2023&rft.volume=11&rft.spage=111566&rft.epage=111580&rft.pages=111566-111580&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2023.3322433&rft_dat=%3Cproquest_ieee_%3E2878510088%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2878510088&rft_id=info:pmid/&rft_ieee_id=10273253&rft_doaj_id=oai_doaj_org_article_254cf39418b14c3484a7a4ea4d40b499&rfr_iscdi=true |