Human activity classification based on sound recognition and residual convolutional neural network
Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound...
Gespeichert in:
Veröffentlicht in: | Automation in construction 2020-06, Vol.114, p.103177, Article 103177 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | 103177 |
container_title | Automation in construction |
container_volume | 114 |
creator | Jung, Minhyuk Chi, Seokho |
description | Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time.
•The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed. |
doi_str_mv | 10.1016/j.autcon.2020.103177 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2440678032</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0926580519307563</els_id><sourcerecordid>2440678032</sourcerecordid><originalsourceid>FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</originalsourceid><addsrcrecordid>eNp9kD1PwzAQhi0EEqXwDxgiMaecndR2FiRUAUWqxAKz5Tg2cmjj4o-i_nuchpnpvt-7exC6xbDAgOl9v5ApKjcsCJAxVWHGztAMc0ZKxht8jmbQEFouOSwv0VUIPQAwoM0Mteu0k0MhVbQHG4-F2soQrLFKRuuGopVBd0V2gktDV3it3OdgTyV5ioPtktwWefnBbdNYyNGgkz-Z-OP81zW6MHIb9M2fnaOP56f31brcvL28rh43paoBYkkpayqjjKmo6hpaGUO5qhvNDG-aFkiHmVZcU9KyVhFDjdGkhorXmLW8W0I1R3eT7t6776RDFL1LPt8TBKlroIxDRXJXPXUp70Lw2oi9tzvpjwKDGGmKXkw0xUhTTDTz2MM0pvMHB6u9CMrqQenOZihRdM7-L_ALN9OB0Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2440678032</pqid></control><display><type>article</type><title>Human activity classification based on sound recognition and residual convolutional neural network</title><source>Elsevier ScienceDirect Journals Complete - AutoHoldings</source><creator>Jung, Minhyuk ; Chi, Seokho</creator><creatorcontrib>Jung, Minhyuk ; Chi, Seokho</creatorcontrib><description>Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time.
•The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.</description><identifier>ISSN: 0926-5805</identifier><identifier>EISSN: 1872-7891</identifier><identifier>DOI: 10.1016/j.autcon.2020.103177</identifier><language>eng</language><publisher>Amsterdam: Elsevier B.V</publisher><subject>Acoustics ; Artificial neural networks ; Classification ; Construction sites ; Convolutional neural network ; Feature extraction ; Filter banks ; Human activity recognition ; Indoor environments ; Machine learning ; Model accuracy ; Moving object recognition ; Neural networks ; Residual neural network ; Sound ; Sound recognition ; Urban environments</subject><ispartof>Automation in construction, 2020-06, Vol.114, p.103177, Article 103177</ispartof><rights>2020 Elsevier B.V.</rights><rights>Copyright Elsevier BV Jun 2020</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</citedby><cites>FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.autcon.2020.103177$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,778,782,3539,27911,27912,45982</link.rule.ids></links><search><creatorcontrib>Jung, Minhyuk</creatorcontrib><creatorcontrib>Chi, Seokho</creatorcontrib><title>Human activity classification based on sound recognition and residual convolutional neural network</title><title>Automation in construction</title><description>Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time.
•The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.</description><subject>Acoustics</subject><subject>Artificial neural networks</subject><subject>Classification</subject><subject>Construction sites</subject><subject>Convolutional neural network</subject><subject>Feature extraction</subject><subject>Filter banks</subject><subject>Human activity recognition</subject><subject>Indoor environments</subject><subject>Machine learning</subject><subject>Model accuracy</subject><subject>Moving object recognition</subject><subject>Neural networks</subject><subject>Residual neural network</subject><subject>Sound</subject><subject>Sound recognition</subject><subject>Urban environments</subject><issn>0926-5805</issn><issn>1872-7891</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNp9kD1PwzAQhi0EEqXwDxgiMaecndR2FiRUAUWqxAKz5Tg2cmjj4o-i_nuchpnpvt-7exC6xbDAgOl9v5ApKjcsCJAxVWHGztAMc0ZKxht8jmbQEFouOSwv0VUIPQAwoM0Mteu0k0MhVbQHG4-F2soQrLFKRuuGopVBd0V2gktDV3it3OdgTyV5ioPtktwWefnBbdNYyNGgkz-Z-OP81zW6MHIb9M2fnaOP56f31brcvL28rh43paoBYkkpayqjjKmo6hpaGUO5qhvNDG-aFkiHmVZcU9KyVhFDjdGkhorXmLW8W0I1R3eT7t6776RDFL1LPt8TBKlroIxDRXJXPXUp70Lw2oi9tzvpjwKDGGmKXkw0xUhTTDTz2MM0pvMHB6u9CMrqQenOZihRdM7-L_ALN9OB0Q</recordid><startdate>202006</startdate><enddate>202006</enddate><creator>Jung, Minhyuk</creator><creator>Chi, Seokho</creator><general>Elsevier B.V</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>202006</creationdate><title>Human activity classification based on sound recognition and residual convolutional neural network</title><author>Jung, Minhyuk ; Chi, Seokho</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c400t-66793fcff36cd963ff68c49e7f899b02d17ec8e62b7bc2f6ffe24038417b8d503</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Acoustics</topic><topic>Artificial neural networks</topic><topic>Classification</topic><topic>Construction sites</topic><topic>Convolutional neural network</topic><topic>Feature extraction</topic><topic>Filter banks</topic><topic>Human activity recognition</topic><topic>Indoor environments</topic><topic>Machine learning</topic><topic>Model accuracy</topic><topic>Moving object recognition</topic><topic>Neural networks</topic><topic>Residual neural network</topic><topic>Sound</topic><topic>Sound recognition</topic><topic>Urban environments</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Jung, Minhyuk</creatorcontrib><creatorcontrib>Chi, Seokho</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Automation in construction</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Jung, Minhyuk</au><au>Chi, Seokho</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Human activity classification based on sound recognition and residual convolutional neural network</atitle><jtitle>Automation in construction</jtitle><date>2020-06</date><risdate>2020</risdate><volume>114</volume><spage>103177</spage><pages>103177-</pages><artnum>103177</artnum><issn>0926-5805</issn><eissn>1872-7891</eissn><abstract>Human activity recognition is crucial for a better understanding of workers in construction sites and people in the built environment. Previous studies have been proposed various ways in which sensing and machine learning techniques can be utilized to collect human activity data automatically. Sound recognition has the potential to be utilized in ways that complement the limitations of the previous methods because sound signals are easy to propagate in indoor environments where many physical obstacles exist, and this method can simultaneously recognize not only sounds from human activities but also sounds from related objects. Therefore, this study develops a sound recognition-based human activity classification model using a residual neural network. A sound data is collected based on ten classes representing people's daily activities in the indoor environment. Then, the features of the sound data were extracted using the Log Mel-filter bank energies method, and a residual neural network model with 34 convolutional layers was trained using the data. The results showed the following: the accuracy of the model was 87.6%, and the Precision score for each class ranged from 76.8% to 92.6%, the Recall scores ranged from 75.8% to 98.6%, and the F1-score ranged from 78.6% to 93.7%. The contribution of this study is to demonstrate that sound recognition can classify people's indoor activities successfully, but this study leaves the limitation that it is based on a monophonic method that only one activity can be classified at a time.
•The human activity classification model was proposed based on a sound recognition method.•The sound dataset for ten human activity classes was developed using open-source video and audio platforms.•A deep residual neural network with 34 convolutional layers was designed to classify sound data converted into 2D spectrograms.•The performance of the classification model by human activity class was evaluated and discussed.</abstract><cop>Amsterdam</cop><pub>Elsevier B.V</pub><doi>10.1016/j.autcon.2020.103177</doi></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0926-5805 |
ispartof | Automation in construction, 2020-06, Vol.114, p.103177, Article 103177 |
issn | 0926-5805 1872-7891 |
language | eng |
recordid | cdi_proquest_journals_2440678032 |
source | Elsevier ScienceDirect Journals Complete - AutoHoldings |
subjects | Acoustics Artificial neural networks Classification Construction sites Convolutional neural network Feature extraction Filter banks Human activity recognition Indoor environments Machine learning Model accuracy Moving object recognition Neural networks Residual neural network Sound Sound recognition Urban environments |
title | Human activity classification based on sound recognition and residual convolutional neural network |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T13%3A30%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Human%20activity%20classification%20based%20on%20sound%20recognition%20and%20residual%20convolutional%20neural%20network&rft.jtitle=Automation%20in%20construction&rft.au=Jung,%20Minhyuk&rft.date=2020-06&rft.volume=114&rft.spage=103177&rft.pages=103177-&rft.artnum=103177&rft.issn=0926-5805&rft.eissn=1872-7891&rft_id=info:doi/10.1016/j.autcon.2020.103177&rft_dat=%3Cproquest_cross%3E2440678032%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2440678032&rft_id=info:pmid/&rft_els_id=S0926580519307563&rfr_iscdi=true |