Human activity classification based on sound recognition and residual convolutional neural network

Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Automation in construction 2020-06, Vol.114, p.103177, Article 103177
Hauptverfasser: Jung, Minhyuk, Chi, Seokho
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 103177
container_title Automation in construction
container_volume 114
creator Jung, Minhyuk
Chi, Seokho
description Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time. •The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.
doi_str_mv 10.1016/j.autcon.2020.103177
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2440678032</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0926580519307563</els_id><sourcerecordid>2440678032</sourcerecordid><originalsourceid>FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</originalsourceid><addsrcrecordid>eNp9kD1PwzAQhi0EEqXwDxgiMaecndR2FiRUAUWqxAKz5Tg2cmjj4o-i_nuchpnpvt-7exC6xbDAgOl9v5ApKjcsCJAxVWHGztAMc0ZKxht8jmbQEFouOSwv0VUIPQAwoM0Mteu0k0MhVbQHG4-F2soQrLFKRuuGopVBd0V2gktDV3it3OdgTyV5ioPtktwWefnBbdNYyNGgkz-Z-OP81zW6MHIb9M2fnaOP56f31brcvL28rh43paoBYkkpayqjjKmo6hpaGUO5qhvNDG-aFkiHmVZcU9KyVhFDjdGkhorXmLW8W0I1R3eT7t6776RDFL1LPt8TBKlroIxDRXJXPXUp70Lw2oi9tzvpjwKDGGmKXkw0xUhTTDTz2MM0pvMHB6u9CMrqQenOZihRdM7-L_ALN9OB0Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2440678032</pqid></control><display><type>article</type><title>Human activity classification based on sound recognition and residual convolutional neural network</title><source>Elsevier ScienceDirect Journals Complete - AutoHoldings</source><creator>Jung, Minhyuk ; Chi, Seokho</creator><creatorcontrib>Jung, Minhyuk ; Chi, Seokho</creatorcontrib><description>Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time. •The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.</description><identifier>ISSN: 0926-5805</identifier><identifier>EISSN: 1872-7891</identifier><identifier>DOI: 10.1016/j.autcon.2020.103177</identifier><language>eng</language><publisher>Amsterdam: Elsevier B.V</publisher><subject>Acoustics ; Artificial neural networks ; Classification ; Construction sites ; Convolutional neural network ; Feature extraction ; Filter banks ; Human activity recognition ; Indoor environments ; Machine learning ; Model accuracy ; Moving object recognition ; Neural networks ; Residual neural network ; Sound ; Sound recognition ; Urban environments</subject><ispartof>Automation in construction, 2020-06, Vol.114, p.103177, Article 103177</ispartof><rights>2020 Elsevier B.V.</rights><rights>Copyright Elsevier BV Jun 2020</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</citedby><cites>FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.autcon.2020.103177$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,778,782,3539,27911,27912,45982</link.rule.ids></links><search><creatorcontrib>Jung, Minhyuk</creatorcontrib><creatorcontrib>Chi, Seokho</creatorcontrib><title>Human activity classification based on sound recognition and residual convolutional neural network</title><title>Automation in construction</title><description>Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time. •The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.</description><subject>Acoustics</subject><subject>Artificial neural networks</subject><subject>Classification</subject><subject>Construction sites</subject><subject>Convolutional neural network</subject><subject>Feature extraction</subject><subject>Filter banks</subject><subject>Human activity recognition</subject><subject>Indoor environments</subject><subject>Machine learning</subject><subject>Model accuracy</subject><subject>Moving object recognition</subject><subject>Neural networks</subject><subject>Residual neural network</subject><subject>Sound</subject><subject>Sound recognition</subject><subject>Urban environments</subject><issn>0926-5805</issn><issn>1872-7891</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNp9kD1PwzAQhi0EEqXwDxgiMaecndR2FiRUAUWqxAKz5Tg2cmjj4o-i_nuchpnpvt-7exC6xbDAgOl9v5ApKjcsCJAxVWHGztAMc0ZKxht8jmbQEFouOSwv0VUIPQAwoM0Mteu0k0MhVbQHG4-F2soQrLFKRuuGopVBd0V2gktDV3it3OdgTyV5ioPtktwWefnBbdNYyNGgkz-Z-OP81zW6MHIb9M2fnaOP56f31brcvL28rh43paoBYkkpayqjjKmo6hpaGUO5qhvNDG-aFkiHmVZcU9KyVhFDjdGkhorXmLW8W0I1R3eT7t6776RDFL1LPt8TBKlroIxDRXJXPXUp70Lw2oi9tzvpjwKDGGmKXkw0xUhTTDTz2MM0pvMHB6u9CMrqQenOZihRdM7-L_ALN9OB0Q</recordid><startdate>202006</startdate><enddate>202006</enddate><creator>Jung, Minhyuk</creator><creator>Chi, Seokho</creator><general>Elsevier B.V</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>202006</creationdate><title>Human activity classification based on sound recognition and residual convolutional neural network</title><author>Jung, Minhyuk ; Chi, Seokho</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Acoustics</topic><topic>Artificial neural networks</topic><topic>Classification</topic><topic>Construction sites</topic><topic>Convolutional neural network</topic><topic>Feature extraction</topic><topic>Filter banks</topic><topic>Human activity recognition</topic><topic>Indoor environments</topic><topic>Machine learning</topic><topic>Model accuracy</topic><topic>Moving object recognition</topic><topic>Neural networks</topic><topic>Residual neural network</topic><topic>Sound</topic><topic>Sound recognition</topic><topic>Urban environments</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Jung, Minhyuk</creatorcontrib><creatorcontrib>Chi, Seokho</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Automation in construction</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Jung, Minhyuk</au><au>Chi, Seokho</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Human activity classification based on sound recognition and residual convolutional neural network</atitle><jtitle>Automation in construction</jtitle><date>2020-06</date><risdate>2020</risdate><volume>114</volume><spage>103177</spage><pages>103177-</pages><artnum>103177</artnum><issn>0926-5805</issn><eissn>1872-7891</eissn><abstract>Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time. •The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.</abstract><cop>Amsterdam</cop><pub>Elsevier B.V</pub><doi>10.1016/j.autcon.2020.103177</doi></addata></record>
fulltext fulltext
identifier ISSN: 0926-5805
ispartof Automation in construction, 2020-06, Vol.114, p.103177, Article 103177
issn 0926-5805
1872-7891
language eng
recordid cdi_proquest_journals_2440678032
source Elsevier ScienceDirect Journals Complete - AutoHoldings
subjects Acoustics
Artificial neural networks
Classification
Construction sites
Convolutional neural network
Feature extraction
Filter banks
Human activity recognition
Indoor environments
Machine learning
Model accuracy
Moving object recognition
Neural networks
Residual neural network
Sound
Sound recognition
Urban environments
title Human activity classification based on sound recognition and residual convolutional neural network
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T13%3A30%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Human%20activity%20classification%20based%20on%20sound%20recognition%20and%20residual%20convolutional%20neural%20network&rft.jtitle=Automation%20in%20construction&rft.au=Jung,%20Minhyuk&rft.date=2020-06&rft.volume=114&rft.spage=103177&rft.pages=103177-&rft.artnum=103177&rft.issn=0926-5805&rft.eissn=1872-7891&rft_id=info:doi/10.1016/j.autcon.2020.103177&rft_dat=%3Cproquest_cross%3E2440678032%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2440678032&rft_id=info:pmid/&rft_els_id=S0926580519307563&rfr_iscdi=true