Revisiting the Effectiveness of 3D Object Recognition Benchmarks

Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2023, Vol.11, p.111566-111580
Hauptverfasser: Song, Hyunsoo, Lee, Seungkyu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 111580
container_issue
container_start_page 111566
container_title IEEE access
container_volume 11
creator Song, Hyunsoo
Lee, Seungkyu
description Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.
doi_str_mv 10.1109/ACCESS.2023.3322433
format Article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_10273253</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10273253</ieee_id><doaj_id>oai_doaj_org_article_254cf39418b14c3484a7a4ea4d40b499</doaj_id><sourcerecordid>2878510088</sourcerecordid><originalsourceid>FETCH-LOGICAL-c359t-649bf7e4ce8f14e61b22f18794e3e99a47f8adf4e145816a000a109ec982d9a43</originalsourceid><addsrcrecordid>eNpNUNFOwjAUXYwmEuQL9GGJz-Dae8faNxFRSUhIQJ-brruFIa64DhL_3uKIoS9tTs8599wTRbcsGTCWyIfReDxZLgc84TAA4BwBLqIOZ0PZhxSGl2fv66jn_SYJRwQozTrR44IOpS-bslrFzZriibVkmvJAFXkfOxvDczzPNwGLF2TcqgpUV8VPVJn1l64__U10ZfXWU-90d6OPl8n7-K0_m79Ox6NZ30Aqm_4QZW4zQkPCMqQhyzm3TGQSCUhKjZkVurBIDNOQTYeIOixHRgpehG_oRtPWt3B6o3Z1Gab_KKdL9Qe4eqV03ZRmS4qnaCxIZCJnaAAF6kwjaSwwyVHK4HXfeu1q970n36iN29dViK-4yETKQj8isKBlmdp5X5P9n8oSdWxetc2rY_Pq1HxQ3bWqkojOFDwDngL8Ap99fT8</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2878510088</pqid></control><display><type>article</type><title>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Song, Hyunsoo ; Lee, Seungkyu</creator><creatorcontrib>Song, Hyunsoo ; Lee, Seungkyu</creatorcontrib><description>Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2023.3322433</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Benchmark ; Benchmark testing ; Benchmarks ; Computer vision ; Data acquisition ; Data augmentation ; dataset ; Datasets ; Effectiveness ; Industrial applications ; Motion perception ; Object recognition ; point cloud ; Point cloud compression ; Scene analysis ; Solid modeling ; Task analysis ; Three dimensional bodies ; Three dimensional models ; Three dimensional motion ; Three-dimensional displays</subject><ispartof>IEEE access, 2023, Vol.11, p.111566-111580</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c359t-649bf7e4ce8f14e61b22f18794e3e99a47f8adf4e145816a000a109ec982d9a43</cites><orcidid>0009-0001-5279-0505 ; 0000-0002-9721-4093</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10273253$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2102,4024,27633,27923,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Song, Hyunsoo</creatorcontrib><creatorcontrib>Lee, Seungkyu</creatorcontrib><title>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</title><title>IEEE access</title><addtitle>Access</addtitle><description>Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.</description><subject>Benchmark</subject><subject>Benchmark testing</subject><subject>Benchmarks</subject><subject>Computer vision</subject><subject>Data acquisition</subject><subject>Data augmentation</subject><subject>dataset</subject><subject>Datasets</subject><subject>Effectiveness</subject><subject>Industrial applications</subject><subject>Motion perception</subject><subject>Object recognition</subject><subject>point cloud</subject><subject>Point cloud compression</subject><subject>Scene analysis</subject><subject>Solid modeling</subject><subject>Task analysis</subject><subject>Three dimensional bodies</subject><subject>Three dimensional models</subject><subject>Three dimensional motion</subject><subject>Three-dimensional displays</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUNFOwjAUXYwmEuQL9GGJz-Dae8faNxFRSUhIQJ-brruFIa64DhL_3uKIoS9tTs8599wTRbcsGTCWyIfReDxZLgc84TAA4BwBLqIOZ0PZhxSGl2fv66jn_SYJRwQozTrR44IOpS-bslrFzZriibVkmvJAFXkfOxvDczzPNwGLF2TcqgpUV8VPVJn1l64__U10ZfXWU-90d6OPl8n7-K0_m79Ox6NZ30Aqm_4QZW4zQkPCMqQhyzm3TGQSCUhKjZkVurBIDNOQTYeIOixHRgpehG_oRtPWt3B6o3Z1Gab_KKdL9Qe4eqV03ZRmS4qnaCxIZCJnaAAF6kwjaSwwyVHK4HXfeu1q970n36iN29dViK-4yETKQj8isKBlmdp5X5P9n8oSdWxetc2rY_Pq1HxQ3bWqkojOFDwDngL8Ap99fT8</recordid><startdate>2023</startdate><enddate>2023</enddate><creator>Song, Hyunsoo</creator><creator>Lee, Seungkyu</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0009-0001-5279-0505</orcidid><orcidid>https://orcid.org/0000-0002-9721-4093</orcidid></search><sort><creationdate>2023</creationdate><title>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</title><author>Song, Hyunsoo ; Lee, Seungkyu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c359t-649bf7e4ce8f14e61b22f18794e3e99a47f8adf4e145816a000a109ec982d9a43</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Benchmark</topic><topic>Benchmark testing</topic><topic>Benchmarks</topic><topic>Computer vision</topic><topic>Data acquisition</topic><topic>Data augmentation</topic><topic>dataset</topic><topic>Datasets</topic><topic>Effectiveness</topic><topic>Industrial applications</topic><topic>Motion perception</topic><topic>Object recognition</topic><topic>point cloud</topic><topic>Point cloud compression</topic><topic>Scene analysis</topic><topic>Solid modeling</topic><topic>Task analysis</topic><topic>Three dimensional bodies</topic><topic>Three dimensional models</topic><topic>Three dimensional motion</topic><topic>Three-dimensional displays</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Song, Hyunsoo</creatorcontrib><creatorcontrib>Lee, Seungkyu</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Song, Hyunsoo</au><au>Lee, Seungkyu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Revisiting the Effectiveness of 3D Object Recognition Benchmarks</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2023</date><risdate>2023</risdate><volume>11</volume><spage>111566</spage><epage>111580</epage><pages>111566-111580</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Recently, 3D computer vision has greatly emerged and become essential topic in both research and industry applications. Yet large scale 3D benchmark like ImageNet is not available for many 3D computer vision tasks such as 3D object recognition, 3D body motion recognition, and 3D scene understanding. Existing 3D benchmarks are not enough in the number of classes and quality of data samples, and reported performances on the datasets are nearly saturated. Furthermore, 3D data obtained with existing 3D sensors are noisy and incomplete causing unreliable evaluation results. In this work, we revisit the effectiveness of existing 3D computer vision benchmarks. We propose to refine and re-organize existing benchmarks to provide cheap and easy access but challenging, effective and reliable evaluation schemes. Our task includes data refinement, class category adjusting, and improved evaluation protocols. Biased benchmark subsets and new challenges are suggested. Our experimental evaluations on ModelNet40, a 3D object recognition benchmark, show that our revised benchmark datasets (MN40-CR and MN20-CB) provide improved indicators for performance comparison and reveals new aspects of existing methods. State-of-the-art 3D object classification and data augmentation methods are evaluated on MN40-CR and MN20-CB. Based on our extensive evaluation, we conclude that existing benchmarks that are carefully re-organized are good alternatives of large scale benchmark which is very expensive to build and difficult to guarantee data quality under immature 3D data acquisition environment. We make our new benchmarks and evaluations public.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2023.3322433</doi><tpages>15</tpages><orcidid>https://orcid.org/0009-0001-5279-0505</orcidid><orcidid>https://orcid.org/0000-0002-9721-4093</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2023, Vol.11, p.111566-111580
issn 2169-3536
2169-3536
language eng
recordid cdi_ieee_primary_10273253
source IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects Benchmark
Benchmark testing
Benchmarks
Computer vision
Data acquisition
Data augmentation
dataset
Datasets
Effectiveness
Industrial applications
Motion perception
Object recognition
point cloud
Point cloud compression
Scene analysis
Solid modeling
Task analysis
Three dimensional bodies
Three dimensional models
Three dimensional motion
Three-dimensional displays
title Revisiting the Effectiveness of 3D Object Recognition Benchmarks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T18%3A36%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Revisiting%20the%20Effectiveness%20of%203D%20Object%20Recognition%20Benchmarks&rft.jtitle=IEEE%20access&rft.au=Song,%20Hyunsoo&rft.date=2023&rft.volume=11&rft.spage=111566&rft.epage=111580&rft.pages=111566-111580&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2023.3322433&rft_dat=%3Cproquest_ieee_%3E2878510088%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2878510088&rft_id=info:pmid/&rft_ieee_id=10273253&rft_doaj_id=oai_doaj_org_article_254cf39418b14c3484a7a4ea4d40b499&rfr_iscdi=true