Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network

As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Sustainability 2024-09, Vol.16 (17), p.7536
Hauptverfasser: Zhou, Xiaotao, Wang, Ning, Hu, Kunrong, Wang, Leiguang, Yu, Chunjiang, Guan, Zhenhua, Hu, Ruiqi, Li, Qiumei, Ye, Longjia
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 17
container_start_page 7536
container_title Sustainability
container_volume 16
creator Zhou, Xiaotao
Wang, Ning
Hu, Kunrong
Wang, Leiguang
Yu, Chunjiang
Guan, Zhenhua
Hu, Ruiqi
Li, Qiumei
Ye, Longjia
description As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.
doi_str_mv 10.3390/su16177536
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3104151874</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3104151874</sourcerecordid><originalsourceid>FETCH-LOGICAL-c184t-f7435cf7a53b8671f17d1d159d2627e884732c057aa302399d210052f7468ac83</originalsourceid><addsrcrecordid>eNpNUFFLwzAQDqLgmHvxFxR8E6K5pmnSx63qFKaCnfgYsjQd3WozkxTx35s5Qe_l7r7v7r7jQ-gcyBWlBbn2A-TAOaP5ERqlhAMGwsjxv_oUTbzfkBiUQgH5CK1fjLbrvg2t7RPbJG_GB-P6ZNYpvcWl27d1Mm9Xq8iXquuSql33KgyRSWbKRzIS1VTemN6bJxPwolo-4mkIpv-5GaFP67Zn6KRRnTeT3zxGr3e3y_IeL57nD-V0gTWILOCGZ5TphitGVyLn0ACvoQZW1GmeciNExmmqCeNKUZLSIuJACEvjXi6UFnSMLg53d85-DPF7ubGD66OkpEAyYCCixBhdHqa0s94708ida9-V-5JA5N5L-ecl_QaG2mSC</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3104151874</pqid></control><display><type>article</type><title>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</title><source>MDPI - Multidisciplinary Digital Publishing Institute</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Zhou, Xiaotao ; Wang, Ning ; Hu, Kunrong ; Wang, Leiguang ; Yu, Chunjiang ; Guan, Zhenhua ; Hu, Ruiqi ; Li, Qiumei ; Ye, Longjia</creator><creatorcontrib>Zhou, Xiaotao ; Wang, Ning ; Hu, Kunrong ; Wang, Leiguang ; Yu, Chunjiang ; Guan, Zhenhua ; Hu, Ruiqi ; Li, Qiumei ; Ye, Longjia</creatorcontrib><description>As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.</description><identifier>ISSN: 2071-1050</identifier><identifier>EISSN: 2071-1050</identifier><identifier>DOI: 10.3390/su16177536</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Accuracy ; Acoustics ; Animals ; Classification ; Datasets ; Deep learning ; Neural networks ; Sound</subject><ispartof>Sustainability, 2024-09, Vol.16 (17), p.7536</ispartof><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c184t-f7435cf7a53b8671f17d1d159d2627e884732c057aa302399d210052f7468ac83</cites><orcidid>0009-0003-9989-5042 ; 0000-0001-8847-4873 ; 0000-0002-6709-4545</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Zhou, Xiaotao</creatorcontrib><creatorcontrib>Wang, Ning</creatorcontrib><creatorcontrib>Hu, Kunrong</creatorcontrib><creatorcontrib>Wang, Leiguang</creatorcontrib><creatorcontrib>Yu, Chunjiang</creatorcontrib><creatorcontrib>Guan, Zhenhua</creatorcontrib><creatorcontrib>Hu, Ruiqi</creatorcontrib><creatorcontrib>Li, Qiumei</creatorcontrib><creatorcontrib>Ye, Longjia</creatorcontrib><title>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</title><title>Sustainability</title><description>As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.</description><subject>Accuracy</subject><subject>Acoustics</subject><subject>Animals</subject><subject>Classification</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Neural networks</subject><subject>Sound</subject><issn>2071-1050</issn><issn>2071-1050</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNpNUFFLwzAQDqLgmHvxFxR8E6K5pmnSx63qFKaCnfgYsjQd3WozkxTx35s5Qe_l7r7v7r7jQ-gcyBWlBbn2A-TAOaP5ERqlhAMGwsjxv_oUTbzfkBiUQgH5CK1fjLbrvg2t7RPbJG_GB-P6ZNYpvcWl27d1Mm9Xq8iXquuSql33KgyRSWbKRzIS1VTemN6bJxPwolo-4mkIpv-5GaFP67Zn6KRRnTeT3zxGr3e3y_IeL57nD-V0gTWILOCGZ5TphitGVyLn0ACvoQZW1GmeciNExmmqCeNKUZLSIuJACEvjXi6UFnSMLg53d85-DPF7ubGD66OkpEAyYCCixBhdHqa0s94708ida9-V-5JA5N5L-ecl_QaG2mSC</recordid><startdate>20240901</startdate><enddate>20240901</enddate><creator>Zhou, Xiaotao</creator><creator>Wang, Ning</creator><creator>Hu, Kunrong</creator><creator>Wang, Leiguang</creator><creator>Yu, Chunjiang</creator><creator>Guan, Zhenhua</creator><creator>Hu, Ruiqi</creator><creator>Li, Qiumei</creator><creator>Ye, Longjia</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>4U-</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><orcidid>https://orcid.org/0009-0003-9989-5042</orcidid><orcidid>https://orcid.org/0000-0001-8847-4873</orcidid><orcidid>https://orcid.org/0000-0002-6709-4545</orcidid></search><sort><creationdate>20240901</creationdate><title>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</title><author>Zhou, Xiaotao ; Wang, Ning ; Hu, Kunrong ; Wang, Leiguang ; Yu, Chunjiang ; Guan, Zhenhua ; Hu, Ruiqi ; Li, Qiumei ; Ye, Longjia</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c184t-f7435cf7a53b8671f17d1d159d2627e884732c057aa302399d210052f7468ac83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Acoustics</topic><topic>Animals</topic><topic>Classification</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Neural networks</topic><topic>Sound</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Xiaotao</creatorcontrib><creatorcontrib>Wang, Ning</creatorcontrib><creatorcontrib>Hu, Kunrong</creatorcontrib><creatorcontrib>Wang, Leiguang</creatorcontrib><creatorcontrib>Yu, Chunjiang</creatorcontrib><creatorcontrib>Guan, Zhenhua</creatorcontrib><creatorcontrib>Hu, Ruiqi</creatorcontrib><creatorcontrib>Li, Qiumei</creatorcontrib><creatorcontrib>Ye, Longjia</creatorcontrib><collection>CrossRef</collection><collection>University Readers</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><jtitle>Sustainability</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhou, Xiaotao</au><au>Wang, Ning</au><au>Hu, Kunrong</au><au>Wang, Leiguang</au><au>Yu, Chunjiang</au><au>Guan, Zhenhua</au><au>Hu, Ruiqi</au><au>Li, Qiumei</au><au>Ye, Longjia</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</atitle><jtitle>Sustainability</jtitle><date>2024-09-01</date><risdate>2024</risdate><volume>16</volume><issue>17</issue><spage>7536</spage><pages>7536-</pages><issn>2071-1050</issn><eissn>2071-1050</eissn><abstract>As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/su16177536</doi><orcidid>https://orcid.org/0009-0003-9989-5042</orcidid><orcidid>https://orcid.org/0000-0001-8847-4873</orcidid><orcidid>https://orcid.org/0000-0002-6709-4545</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2071-1050
ispartof Sustainability, 2024-09, Vol.16 (17), p.7536
issn 2071-1050
2071-1050
language eng
recordid cdi_proquest_journals_3104151874
source MDPI - Multidisciplinary Digital Publishing Institute; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects Accuracy
Acoustics
Animals
Classification
Datasets
Deep learning
Neural networks
Sound
title Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-03T17%3A00%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Recognition%20of%20Western%20Black-Crested%20Gibbon%20Call%20Signatures%20Based%20on%20SA_DenseNet-LSTM-Attention%20Network&rft.jtitle=Sustainability&rft.au=Zhou,%20Xiaotao&rft.date=2024-09-01&rft.volume=16&rft.issue=17&rft.spage=7536&rft.pages=7536-&rft.issn=2071-1050&rft.eissn=2071-1050&rft_id=info:doi/10.3390/su16177536&rft_dat=%3Cproquest_cross%3E3104151874%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3104151874&rft_id=info:pmid/&rfr_iscdi=true