Triple-Classification of Respiratory Sounds Using Optimized S-Transform and Deep Residual Networks
Digital respiratory sounds provide valuable information for telemedicine and smart diagnosis in an non-invasive way of pathological detection. As the typical continuous abnormal respiratory sound, wheeze is clinically correlated with asthma or chronic obstructive lung diseases. Meanwhile, the discon...
Gespeichert in:
Veröffentlicht in: | IEEE access 2019, Vol.7, p.32845-32852 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 32852 |
---|---|
container_issue | |
container_start_page | 32845 |
container_title | IEEE access |
container_volume | 7 |
creator | Chen, Hai Yuan, Xiaochen Pei, Zhiyuan Li, Mianjie Li, Jianqing |
description | Digital respiratory sounds provide valuable information for telemedicine and smart diagnosis in an non-invasive way of pathological detection. As the typical continuous abnormal respiratory sound, wheeze is clinically correlated with asthma or chronic obstructive lung diseases. Meanwhile, the discontinuous adventitious crackle is clinically correlated with pneumonia, bronchitis, and so on. The detection and classification of both attract many studies for decades. However, due to the contained artifacts and constrained feature extraction methods, the reliability and accuracy of the classification of wheeze, crackle, and normal sounds need significant improvement. In this paper, we propose a novel method for the identification of wheeze, crackle, and normal sounds using the optimized S-transform (OST) and deep residual networks (ResNets). First, the raw respiratory sound is processed by the proposed OST. Then, the spectrogram of OST is rescaled for the Resnet. After the feature learning and classification are fulfilled by the ResNet, the classes of respiratory sounds are recognized. Because the proposed OST highlights the features of wheeze, crackle, and respiratory sounds, and the deep residual learning generates discriminative features for better recognition, this proposed method provides reliable access for respiratory disease-related telemedicine and E-health diagnosis. The experimental results show that the proposed OST and ResNet is excellent for the multi-classification of respiratory sounds with the accuracy , sensitivity , and specificity up to 98.79%, 96.27%, and 100%, respectively. The comparison results of the triple-classification of respiratory sounds indicate that the proposed method outperforms the deep-learning-based ensembling convolutional neural network (CNN) by 3.23% and the empirical mode decomposition-based artificial neural network (ANN) by 4.63%, respectively. |
doi_str_mv | 10.1109/ACCESS.2019.2903859 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_ACCESS_2019_2903859</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>8663379</ieee_id><doaj_id>oai_doaj_org_article_bb3e783c780a422eaee0ca566c166980</doaj_id><sourcerecordid>2455612244</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-80af1389b9f76cc05cea8f9679db1f520a234af1fe488e7c26b84b5e47baea493</originalsourceid><addsrcrecordid>eNpNUU1r3DAQNaWFhjS_IBdBz97q29IxOEkTCA10N2cxlkdBW6_lSl5K-uvrjUPoXGZ4vI-BV1WXjG4Yo_bbVdvebLcbTpndcEuFUfZDdcaZtrVQQn_87_5cXZSyp8uYBVLNWdXtcpwGrNsBSokhephjGkkK5CeWKWaYU34h23Qc-0KeShyfyeM0x0P8iz3Z1rsMYwkpHwiMPblGnE662B9hID9w_pPyr_Kl-hRgKHjxts-rp9ubXXtXPzx-v2-vHmovqZlrQyEwYWxnQ6O9p8ojmGB1Y_uOBcUpcCEXSkBpDDae687ITqFsOkCQVpxX96tvn2DvphwPkF9cguhegZSfHeQ5-gFd1wlsjPDNEio5R0CkHpTWnmltDV28vq5eU06_j1hmt0_HPC7vOy6V0oxzKReWWFk-p1IyhvdURt2pG7d2407duLduFtXlqoqI-K4wWgvRWPEPnCuLhQ</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2455612244</pqid></control><display><type>article</type><title>Triple-Classification of Respiratory Sounds Using Optimized S-Transform and Deep Residual Networks</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Chen, Hai ; Yuan, Xiaochen ; Pei, Zhiyuan ; Li, Mianjie ; Li, Jianqing</creator><creatorcontrib>Chen, Hai ; Yuan, Xiaochen ; Pei, Zhiyuan ; Li, Mianjie ; Li, Jianqing</creatorcontrib><description>Digital respiratory sounds provide valuable information for telemedicine and smart diagnosis in an non-invasive way of pathological detection. As the typical continuous abnormal respiratory sound, wheeze is clinically correlated with asthma or chronic obstructive lung diseases. Meanwhile, the discontinuous adventitious crackle is clinically correlated with pneumonia, bronchitis, and so on. The detection and classification of both attract many studies for decades. However, due to the contained artifacts and constrained feature extraction methods, the reliability and accuracy of the classification of wheeze, crackle, and normal sounds need significant improvement. In this paper, we propose a novel method for the identification of wheeze, crackle, and normal sounds using the optimized S-transform (OST) and deep residual networks (ResNets). First, the raw respiratory sound is processed by the proposed OST. Then, the spectrogram of OST is rescaled for the Resnet. After the feature learning and classification are fulfilled by the ResNet, the classes of respiratory sounds are recognized. Because the proposed OST highlights the features of wheeze, crackle, and respiratory sounds, and the deep residual learning generates discriminative features for better recognition, this proposed method provides reliable access for respiratory disease-related telemedicine and E-health diagnosis. The experimental results show that the proposed OST and ResNet is excellent for the multi-classification of respiratory sounds with the accuracy , sensitivity , and specificity up to 98.79%, 96.27%, and 100%, respectively. The comparison results of the triple-classification of respiratory sounds indicate that the proposed method outperforms the deep-learning-based ensembling convolutional neural network (CNN) by 3.23% and the empirical mode decomposition-based artificial neural network (ANN) by 4.63%, respectively.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2019.2903859</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Acoustics ; Artificial neural networks ; Asthma ; Classification ; crackle and wheeze detection ; Deep residual networks (ResNet) ; Diagnosis ; Diseases ; Feature extraction ; Feature recognition ; Learning theory ; Lung ; Machine learning ; Neural networks ; optimized S-transform (OST) ; Respiratory diseases ; respiratory sounds classification ; Sound ; Spectrogram ; Telemedicine ; Time-frequency analysis ; Training ; Transformations (mathematics) ; Transforms</subject><ispartof>IEEE access, 2019, Vol.7, p.32845-32852</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2019</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-80af1389b9f76cc05cea8f9679db1f520a234af1fe488e7c26b84b5e47baea493</citedby><cites>FETCH-LOGICAL-c408t-80af1389b9f76cc05cea8f9679db1f520a234af1fe488e7c26b84b5e47baea493</cites><orcidid>0000-0002-7490-6695 ; 0000-0002-6768-1483</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/8663379$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>315,781,785,865,2103,4025,27638,27928,27929,27930,54938</link.rule.ids></links><search><creatorcontrib>Chen, Hai</creatorcontrib><creatorcontrib>Yuan, Xiaochen</creatorcontrib><creatorcontrib>Pei, Zhiyuan</creatorcontrib><creatorcontrib>Li, Mianjie</creatorcontrib><creatorcontrib>Li, Jianqing</creatorcontrib><title>Triple-Classification of Respiratory Sounds Using Optimized S-Transform and Deep Residual Networks</title><title>IEEE access</title><addtitle>Access</addtitle><description>Digital respiratory sounds provide valuable information for telemedicine and smart diagnosis in an non-invasive way of pathological detection. As the typical continuous abnormal respiratory sound, wheeze is clinically correlated with asthma or chronic obstructive lung diseases. Meanwhile, the discontinuous adventitious crackle is clinically correlated with pneumonia, bronchitis, and so on. The detection and classification of both attract many studies for decades. However, due to the contained artifacts and constrained feature extraction methods, the reliability and accuracy of the classification of wheeze, crackle, and normal sounds need significant improvement. In this paper, we propose a novel method for the identification of wheeze, crackle, and normal sounds using the optimized S-transform (OST) and deep residual networks (ResNets). First, the raw respiratory sound is processed by the proposed OST. Then, the spectrogram of OST is rescaled for the Resnet. After the feature learning and classification are fulfilled by the ResNet, the classes of respiratory sounds are recognized. Because the proposed OST highlights the features of wheeze, crackle, and respiratory sounds, and the deep residual learning generates discriminative features for better recognition, this proposed method provides reliable access for respiratory disease-related telemedicine and E-health diagnosis. The experimental results show that the proposed OST and ResNet is excellent for the multi-classification of respiratory sounds with the accuracy , sensitivity , and specificity up to 98.79%, 96.27%, and 100%, respectively. The comparison results of the triple-classification of respiratory sounds indicate that the proposed method outperforms the deep-learning-based ensembling convolutional neural network (CNN) by 3.23% and the empirical mode decomposition-based artificial neural network (ANN) by 4.63%, respectively.</description><subject>Acoustics</subject><subject>Artificial neural networks</subject><subject>Asthma</subject><subject>Classification</subject><subject>crackle and wheeze detection</subject><subject>Deep residual networks (ResNet)</subject><subject>Diagnosis</subject><subject>Diseases</subject><subject>Feature extraction</subject><subject>Feature recognition</subject><subject>Learning theory</subject><subject>Lung</subject><subject>Machine learning</subject><subject>Neural networks</subject><subject>optimized S-transform (OST)</subject><subject>Respiratory diseases</subject><subject>respiratory sounds classification</subject><subject>Sound</subject><subject>Spectrogram</subject><subject>Telemedicine</subject><subject>Time-frequency analysis</subject><subject>Training</subject><subject>Transformations (mathematics)</subject><subject>Transforms</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUU1r3DAQNaWFhjS_IBdBz97q29IxOEkTCA10N2cxlkdBW6_lSl5K-uvrjUPoXGZ4vI-BV1WXjG4Yo_bbVdvebLcbTpndcEuFUfZDdcaZtrVQQn_87_5cXZSyp8uYBVLNWdXtcpwGrNsBSokhephjGkkK5CeWKWaYU34h23Qc-0KeShyfyeM0x0P8iz3Z1rsMYwkpHwiMPblGnE662B9hID9w_pPyr_Kl-hRgKHjxts-rp9ubXXtXPzx-v2-vHmovqZlrQyEwYWxnQ6O9p8ojmGB1Y_uOBcUpcCEXSkBpDDae687ITqFsOkCQVpxX96tvn2DvphwPkF9cguhegZSfHeQ5-gFd1wlsjPDNEio5R0CkHpTWnmltDV28vq5eU06_j1hmt0_HPC7vOy6V0oxzKReWWFk-p1IyhvdURt2pG7d2407duLduFtXlqoqI-K4wWgvRWPEPnCuLhQ</recordid><startdate>2019</startdate><enddate>2019</enddate><creator>Chen, Hai</creator><creator>Yuan, Xiaochen</creator><creator>Pei, Zhiyuan</creator><creator>Li, Mianjie</creator><creator>Li, Jianqing</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7490-6695</orcidid><orcidid>https://orcid.org/0000-0002-6768-1483</orcidid></search><sort><creationdate>2019</creationdate><title>Triple-Classification of Respiratory Sounds Using Optimized S-Transform and Deep Residual Networks</title><author>Chen, Hai ; Yuan, Xiaochen ; Pei, Zhiyuan ; Li, Mianjie ; Li, Jianqing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-80af1389b9f76cc05cea8f9679db1f520a234af1fe488e7c26b84b5e47baea493</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Acoustics</topic><topic>Artificial neural networks</topic><topic>Asthma</topic><topic>Classification</topic><topic>crackle and wheeze detection</topic><topic>Deep residual networks (ResNet)</topic><topic>Diagnosis</topic><topic>Diseases</topic><topic>Feature extraction</topic><topic>Feature recognition</topic><topic>Learning theory</topic><topic>Lung</topic><topic>Machine learning</topic><topic>Neural networks</topic><topic>optimized S-transform (OST)</topic><topic>Respiratory diseases</topic><topic>respiratory sounds classification</topic><topic>Sound</topic><topic>Spectrogram</topic><topic>Telemedicine</topic><topic>Time-frequency analysis</topic><topic>Training</topic><topic>Transformations (mathematics)</topic><topic>Transforms</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Chen, Hai</creatorcontrib><creatorcontrib>Yuan, Xiaochen</creatorcontrib><creatorcontrib>Pei, Zhiyuan</creatorcontrib><creatorcontrib>Li, Mianjie</creatorcontrib><creatorcontrib>Li, Jianqing</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chen, Hai</au><au>Yuan, Xiaochen</au><au>Pei, Zhiyuan</au><au>Li, Mianjie</au><au>Li, Jianqing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Triple-Classification of Respiratory Sounds Using Optimized S-Transform and Deep Residual Networks</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2019</date><risdate>2019</risdate><volume>7</volume><spage>32845</spage><epage>32852</epage><pages>32845-32852</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Digital respiratory sounds provide valuable information for telemedicine and smart diagnosis in an non-invasive way of pathological detection. As the typical continuous abnormal respiratory sound, wheeze is clinically correlated with asthma or chronic obstructive lung diseases. Meanwhile, the discontinuous adventitious crackle is clinically correlated with pneumonia, bronchitis, and so on. The detection and classification of both attract many studies for decades. However, due to the contained artifacts and constrained feature extraction methods, the reliability and accuracy of the classification of wheeze, crackle, and normal sounds need significant improvement. In this paper, we propose a novel method for the identification of wheeze, crackle, and normal sounds using the optimized S-transform (OST) and deep residual networks (ResNets). First, the raw respiratory sound is processed by the proposed OST. Then, the spectrogram of OST is rescaled for the Resnet. After the feature learning and classification are fulfilled by the ResNet, the classes of respiratory sounds are recognized. Because the proposed OST highlights the features of wheeze, crackle, and respiratory sounds, and the deep residual learning generates discriminative features for better recognition, this proposed method provides reliable access for respiratory disease-related telemedicine and E-health diagnosis. The experimental results show that the proposed OST and ResNet is excellent for the multi-classification of respiratory sounds with the accuracy , sensitivity , and specificity up to 98.79%, 96.27%, and 100%, respectively. The comparison results of the triple-classification of respiratory sounds indicate that the proposed method outperforms the deep-learning-based ensembling convolutional neural network (CNN) by 3.23% and the empirical mode decomposition-based artificial neural network (ANN) by 4.63%, respectively.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2019.2903859</doi><tpages>8</tpages><orcidid>https://orcid.org/0000-0002-7490-6695</orcidid><orcidid>https://orcid.org/0000-0002-6768-1483</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2019, Vol.7, p.32845-32852 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_crossref_primary_10_1109_ACCESS_2019_2903859 |
source | IEEE Open Access Journals; DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals |
subjects | Acoustics Artificial neural networks Asthma Classification crackle and wheeze detection Deep residual networks (ResNet) Diagnosis Diseases Feature extraction Feature recognition Learning theory Lung Machine learning Neural networks optimized S-transform (OST) Respiratory diseases respiratory sounds classification Sound Spectrogram Telemedicine Time-frequency analysis Training Transformations (mathematics) Transforms |
title | Triple-Classification of Respiratory Sounds Using Optimized S-Transform and Deep Residual Networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-11T14%3A36%3A05IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Triple-Classification%20of%20Respiratory%20Sounds%20Using%20Optimized%20S-Transform%20and%20Deep%20Residual%20Networks&rft.jtitle=IEEE%20access&rft.au=Chen,%20Hai&rft.date=2019&rft.volume=7&rft.spage=32845&rft.epage=32852&rft.pages=32845-32852&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2019.2903859&rft_dat=%3Cproquest_cross%3E2455612244%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2455612244&rft_id=info:pmid/&rft_ieee_id=8663379&rft_doaj_id=oai_doaj_org_article_bb3e783c780a422eaee0ca566c166980&rfr_iscdi=true |