PointNet-Transformer Fusion Network for In-Cabin Occupancy Monitoring With mm-Wave Radar

Due to the irregular distribution of 3-D point clouds and the random movement of the passengers, it is difficult to accurately and rapidly determine the passenger occupancy based on the multiple-input and multiple-output (MIMO) radar. In this study, we present a lightweight neural network, dubbed Po...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE sensors journal 2024-02, Vol.24 (4), p.5370-5382
Hauptverfasser: Xiao, Zhiqiang, Ye, Kuntao, Cui, Guolong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 5382
container_issue 4
container_start_page 5370
container_title IEEE sensors journal
container_volume 24
creator Xiao, Zhiqiang
Ye, Kuntao
Cui, Guolong
description Due to the irregular distribution of 3-D point clouds and the random movement of the passengers, it is difficult to accurately and rapidly determine the passenger occupancy based on the multiple-input and multiple-output (MIMO) radar. In this study, we present a lightweight neural network, dubbed PointNet and transformer fusion neural network (PTFNet), that fuses the PointNet and transformer to quickly and accurately detect the occupancy of three rear seats and two footwells. PTFNet adapts an input transform block to encode the 3-D point cloud directly, which makes it more lightweight than conventional voxelization, heatmap, and PointNet based algorithms. Meanwhile, a cross-attention mechanism block inspired by transformer is employed to efficiently extract features from 3-D point clouds, resulting in an improvement in detection accuracy. An in-cabin occupancy monitoring system (OMS) is implemented in a real vehicle to obtain 3-D point cloud datasets, which are then used to evaluate the detection performance of PTFNet. Our datasets cover both the rear seat area and the footwell area. The footwell space is considered for the first time in the point cloud datasets. The experimental results demonstrate that PTFNet outperforms other popular approaches in terms of detection accuracy by at least 1.6%, while the consumption of memory and running time are reduced by at least 44% and 34%, respectively.
doi_str_mv 10.1109/JSEN.2023.3347893
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_journals_2926268032</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10381624</ieee_id><sourcerecordid>2926268032</sourcerecordid><originalsourceid>FETCH-LOGICAL-c246t-4c48690293a54b08516c1cb640ad694ebe7466a6bb493e037e46cd1b884eba7f3</originalsourceid><addsrcrecordid>eNpNkEtPAjEUhRujiYj-ABMXTVwP9jV9LA1BxSAYxcBu0ikdLUqL7YyGf-9MYOHq3px7zj3JB8AlRgOMkbp5fB1NBwQROqCUCanoEejhPJcZFkwedztFGaNieQrOUlojhJXIRQ8sn4Pz9dTW2Txqn6oQNzbCuya54GEr_4b4CVsVjn021KXzcGZMs9Xe7OBT8K4O0fl3uHD1B9xssoX-sfBFr3Q8ByeV_kr24jD74O1uNB8-ZJPZ_Xh4O8kMYbzOmGGSK0QU1TkrkcwxN9iUnCG94orZ0grGueZlyRS1iArLuFnhUsr2pkVF--B6_3cbw3djU12sQxN9W1kQRTjhElHSuvDeZWJIKdqq2Ea30XFXYFR0AIsOYNEBLA4A28zVPuOstf_8VGJOGP0DPdpsMA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2926268032</pqid></control><display><type>article</type><title>PointNet-Transformer Fusion Network for In-Cabin Occupancy Monitoring With mm-Wave Radar</title><source>IEEE Electronic Library (IEL)</source><creator>Xiao, Zhiqiang ; Ye, Kuntao ; Cui, Guolong</creator><creatorcontrib>Xiao, Zhiqiang ; Ye, Kuntao ; Cui, Guolong</creatorcontrib><description>Due to the irregular distribution of 3-D point clouds and the random movement of the passengers, it is difficult to accurately and rapidly determine the passenger occupancy based on the multiple-input and multiple-output (MIMO) radar. In this study, we present a lightweight neural network, dubbed PointNet and transformer fusion neural network (PTFNet), that fuses the PointNet and transformer to quickly and accurately detect the occupancy of three rear seats and two footwells. PTFNet adapts an input transform block to encode the 3-D point cloud directly, which makes it more lightweight than conventional voxelization, heatmap, and PointNet based algorithms. Meanwhile, a cross-attention mechanism block inspired by transformer is employed to efficiently extract features from 3-D point clouds, resulting in an improvement in detection accuracy. An in-cabin occupancy monitoring system (OMS) is implemented in a real vehicle to obtain 3-D point cloud datasets, which are then used to evaluate the detection performance of PTFNet. Our datasets cover both the rear seat area and the footwell area. The footwell space is considered for the first time in the point cloud datasets. The experimental results demonstrate that PTFNet outperforms other popular approaches in terms of detection accuracy by at least 1.6%, while the consumption of memory and running time are reduced by at least 44% and 34%, respectively.</description><identifier>ISSN: 1530-437X</identifier><identifier>EISSN: 1558-1748</identifier><identifier>DOI: 10.1109/JSEN.2023.3347893</identifier><identifier>CODEN: ISJEAZ</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>3-D point cloud ; Algorithms ; Antenna arrays ; Datasets ; Lightweight ; Millimeter waves ; Monitoring ; multiple-input and multiple-output (MIMO) millimeter-wave (mm-Wave) radar sensor ; Neural networks ; Occupancy ; passenger occupancy detection ; Point cloud compression ; PointNet ; Radar ; Radar antennas ; Seats ; Sensors ; Three dimensional models ; Three-dimensional displays ; transformer ; Transformers ; Transmitting antennas</subject><ispartof>IEEE sensors journal, 2024-02, Vol.24 (4), p.5370-5382</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c246t-4c48690293a54b08516c1cb640ad694ebe7466a6bb493e037e46cd1b884eba7f3</cites><orcidid>0000-0001-5780-6310 ; 0000-0002-0360-9925 ; 0000-0001-5707-6311</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10381624$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10381624$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Xiao, Zhiqiang</creatorcontrib><creatorcontrib>Ye, Kuntao</creatorcontrib><creatorcontrib>Cui, Guolong</creatorcontrib><title>PointNet-Transformer Fusion Network for In-Cabin Occupancy Monitoring With mm-Wave Radar</title><title>IEEE sensors journal</title><addtitle>JSEN</addtitle><description>Due to the irregular distribution of 3-D point clouds and the random movement of the passengers, it is difficult to accurately and rapidly determine the passenger occupancy based on the multiple-input and multiple-output (MIMO) radar. In this study, we present a lightweight neural network, dubbed PointNet and transformer fusion neural network (PTFNet), that fuses the PointNet and transformer to quickly and accurately detect the occupancy of three rear seats and two footwells. PTFNet adapts an input transform block to encode the 3-D point cloud directly, which makes it more lightweight than conventional voxelization, heatmap, and PointNet based algorithms. Meanwhile, a cross-attention mechanism block inspired by transformer is employed to efficiently extract features from 3-D point clouds, resulting in an improvement in detection accuracy. An in-cabin occupancy monitoring system (OMS) is implemented in a real vehicle to obtain 3-D point cloud datasets, which are then used to evaluate the detection performance of PTFNet. Our datasets cover both the rear seat area and the footwell area. The footwell space is considered for the first time in the point cloud datasets. The experimental results demonstrate that PTFNet outperforms other popular approaches in terms of detection accuracy by at least 1.6%, while the consumption of memory and running time are reduced by at least 44% and 34%, respectively.</description><subject>3-D point cloud</subject><subject>Algorithms</subject><subject>Antenna arrays</subject><subject>Datasets</subject><subject>Lightweight</subject><subject>Millimeter waves</subject><subject>Monitoring</subject><subject>multiple-input and multiple-output (MIMO) millimeter-wave (mm-Wave) radar sensor</subject><subject>Neural networks</subject><subject>Occupancy</subject><subject>passenger occupancy detection</subject><subject>Point cloud compression</subject><subject>PointNet</subject><subject>Radar</subject><subject>Radar antennas</subject><subject>Seats</subject><subject>Sensors</subject><subject>Three dimensional models</subject><subject>Three-dimensional displays</subject><subject>transformer</subject><subject>Transformers</subject><subject>Transmitting antennas</subject><issn>1530-437X</issn><issn>1558-1748</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpNkEtPAjEUhRujiYj-ABMXTVwP9jV9LA1BxSAYxcBu0ikdLUqL7YyGf-9MYOHq3px7zj3JB8AlRgOMkbp5fB1NBwQROqCUCanoEejhPJcZFkwedztFGaNieQrOUlojhJXIRQ8sn4Pz9dTW2Txqn6oQNzbCuya54GEr_4b4CVsVjn021KXzcGZMs9Xe7OBT8K4O0fl3uHD1B9xssoX-sfBFr3Q8ByeV_kr24jD74O1uNB8-ZJPZ_Xh4O8kMYbzOmGGSK0QU1TkrkcwxN9iUnCG94orZ0grGueZlyRS1iArLuFnhUsr2pkVF--B6_3cbw3djU12sQxN9W1kQRTjhElHSuvDeZWJIKdqq2Ea30XFXYFR0AIsOYNEBLA4A28zVPuOstf_8VGJOGP0DPdpsMA</recordid><startdate>20240215</startdate><enddate>20240215</enddate><creator>Xiao, Zhiqiang</creator><creator>Ye, Kuntao</creator><creator>Cui, Guolong</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>7U5</scope><scope>8FD</scope><scope>L7M</scope><orcidid>https://orcid.org/0000-0001-5780-6310</orcidid><orcidid>https://orcid.org/0000-0002-0360-9925</orcidid><orcidid>https://orcid.org/0000-0001-5707-6311</orcidid></search><sort><creationdate>20240215</creationdate><title>PointNet-Transformer Fusion Network for In-Cabin Occupancy Monitoring With mm-Wave Radar</title><author>Xiao, Zhiqiang ; Ye, Kuntao ; Cui, Guolong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c246t-4c48690293a54b08516c1cb640ad694ebe7466a6bb493e037e46cd1b884eba7f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>3-D point cloud</topic><topic>Algorithms</topic><topic>Antenna arrays</topic><topic>Datasets</topic><topic>Lightweight</topic><topic>Millimeter waves</topic><topic>Monitoring</topic><topic>multiple-input and multiple-output (MIMO) millimeter-wave (mm-Wave) radar sensor</topic><topic>Neural networks</topic><topic>Occupancy</topic><topic>passenger occupancy detection</topic><topic>Point cloud compression</topic><topic>PointNet</topic><topic>Radar</topic><topic>Radar antennas</topic><topic>Seats</topic><topic>Sensors</topic><topic>Three dimensional models</topic><topic>Three-dimensional displays</topic><topic>transformer</topic><topic>Transformers</topic><topic>Transmitting antennas</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Xiao, Zhiqiang</creatorcontrib><creatorcontrib>Ye, Kuntao</creatorcontrib><creatorcontrib>Cui, Guolong</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>Technology Research Database</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>IEEE sensors journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Xiao, Zhiqiang</au><au>Ye, Kuntao</au><au>Cui, Guolong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>PointNet-Transformer Fusion Network for In-Cabin Occupancy Monitoring With mm-Wave Radar</atitle><jtitle>IEEE sensors journal</jtitle><stitle>JSEN</stitle><date>2024-02-15</date><risdate>2024</risdate><volume>24</volume><issue>4</issue><spage>5370</spage><epage>5382</epage><pages>5370-5382</pages><issn>1530-437X</issn><eissn>1558-1748</eissn><coden>ISJEAZ</coden><abstract>Due to the irregular distribution of 3-D point clouds and the random movement of the passengers, it is difficult to accurately and rapidly determine the passenger occupancy based on the multiple-input and multiple-output (MIMO) radar. In this study, we present a lightweight neural network, dubbed PointNet and transformer fusion neural network (PTFNet), that fuses the PointNet and transformer to quickly and accurately detect the occupancy of three rear seats and two footwells. PTFNet adapts an input transform block to encode the 3-D point cloud directly, which makes it more lightweight than conventional voxelization, heatmap, and PointNet based algorithms. Meanwhile, a cross-attention mechanism block inspired by transformer is employed to efficiently extract features from 3-D point clouds, resulting in an improvement in detection accuracy. An in-cabin occupancy monitoring system (OMS) is implemented in a real vehicle to obtain 3-D point cloud datasets, which are then used to evaluate the detection performance of PTFNet. Our datasets cover both the rear seat area and the footwell area. The footwell space is considered for the first time in the point cloud datasets. The experimental results demonstrate that PTFNet outperforms other popular approaches in terms of detection accuracy by at least 1.6%, while the consumption of memory and running time are reduced by at least 44% and 34%, respectively.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/JSEN.2023.3347893</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0001-5780-6310</orcidid><orcidid>https://orcid.org/0000-0002-0360-9925</orcidid><orcidid>https://orcid.org/0000-0001-5707-6311</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1530-437X
ispartof IEEE sensors journal, 2024-02, Vol.24 (4), p.5370-5382
issn 1530-437X
1558-1748
language eng
recordid cdi_proquest_journals_2926268032
source IEEE Electronic Library (IEL)
subjects 3-D point cloud
Algorithms
Antenna arrays
Datasets
Lightweight
Millimeter waves
Monitoring
multiple-input and multiple-output (MIMO) millimeter-wave (mm-Wave) radar sensor
Neural networks
Occupancy
passenger occupancy detection
Point cloud compression
PointNet
Radar
Radar antennas
Seats
Sensors
Three dimensional models
Three-dimensional displays
transformer
Transformers
Transmitting antennas
title PointNet-Transformer Fusion Network for In-Cabin Occupancy Monitoring With mm-Wave Radar
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T01%3A40%3A50IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=PointNet-Transformer%20Fusion%20Network%20for%20In-Cabin%20Occupancy%20Monitoring%20With%20mm-Wave%20Radar&rft.jtitle=IEEE%20sensors%20journal&rft.au=Xiao,%20Zhiqiang&rft.date=2024-02-15&rft.volume=24&rft.issue=4&rft.spage=5370&rft.epage=5382&rft.pages=5370-5382&rft.issn=1530-437X&rft.eissn=1558-1748&rft.coden=ISJEAZ&rft_id=info:doi/10.1109/JSEN.2023.3347893&rft_dat=%3Cproquest_RIE%3E2926268032%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2926268032&rft_id=info:pmid/&rft_ieee_id=10381624&rfr_iscdi=true