Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network
As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation....
Gespeichert in:
Veröffentlicht in: | Sustainability 2024-09, Vol.16 (17), p.7536 |
---|---|
Hauptverfasser: | , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 17 |
container_start_page | 7536 |
container_title | Sustainability |
container_volume | 16 |
creator | Zhou, Xiaotao Wang, Ning Hu, Kunrong Wang, Leiguang Yu, Chunjiang Guan, Zhenhua Hu, Ruiqi Li, Qiumei Ye, Longjia |
description | As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations. |
doi_str_mv | 10.3390/su16177536 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3104151874</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3104151874</sourcerecordid><originalsourceid>FETCH-LOGICAL-c184t-f7435cf7a53b8671f17d1d159d2627e884732c057aa302399d210052f7468ac83</originalsourceid><addsrcrecordid>eNpNUFFLwzAQDqLgmHvxFxR8E6K5pmnSx63qFKaCnfgYsjQd3WozkxTx35s5Qe_l7r7v7r7jQ-gcyBWlBbn2A-TAOaP5ERqlhAMGwsjxv_oUTbzfkBiUQgH5CK1fjLbrvg2t7RPbJG_GB-P6ZNYpvcWl27d1Mm9Xq8iXquuSql33KgyRSWbKRzIS1VTemN6bJxPwolo-4mkIpv-5GaFP67Zn6KRRnTeT3zxGr3e3y_IeL57nD-V0gTWILOCGZ5TphitGVyLn0ACvoQZW1GmeciNExmmqCeNKUZLSIuJACEvjXi6UFnSMLg53d85-DPF7ubGD66OkpEAyYCCixBhdHqa0s94708ida9-V-5JA5N5L-ecl_QaG2mSC</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3104151874</pqid></control><display><type>article</type><title>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</title><source>MDPI - Multidisciplinary Digital Publishing Institute</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Zhou, Xiaotao ; Wang, Ning ; Hu, Kunrong ; Wang, Leiguang ; Yu, Chunjiang ; Guan, Zhenhua ; Hu, Ruiqi ; Li, Qiumei ; Ye, Longjia</creator><creatorcontrib>Zhou, Xiaotao ; Wang, Ning ; Hu, Kunrong ; Wang, Leiguang ; Yu, Chunjiang ; Guan, Zhenhua ; Hu, Ruiqi ; Li, Qiumei ; Ye, Longjia</creatorcontrib><description>As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.</description><identifier>ISSN: 2071-1050</identifier><identifier>EISSN: 2071-1050</identifier><identifier>DOI: 10.3390/su16177536</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Accuracy ; Acoustics ; Animals ; Classification ; Datasets ; Deep learning ; Neural networks ; Sound</subject><ispartof>Sustainability, 2024-09, Vol.16 (17), p.7536</ispartof><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c184t-f7435cf7a53b8671f17d1d159d2627e884732c057aa302399d210052f7468ac83</cites><orcidid>0009-0003-9989-5042 ; 0000-0001-8847-4873 ; 0000-0002-6709-4545</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Zhou, Xiaotao</creatorcontrib><creatorcontrib>Wang, Ning</creatorcontrib><creatorcontrib>Hu, Kunrong</creatorcontrib><creatorcontrib>Wang, Leiguang</creatorcontrib><creatorcontrib>Yu, Chunjiang</creatorcontrib><creatorcontrib>Guan, Zhenhua</creatorcontrib><creatorcontrib>Hu, Ruiqi</creatorcontrib><creatorcontrib>Li, Qiumei</creatorcontrib><creatorcontrib>Ye, Longjia</creatorcontrib><title>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</title><title>Sustainability</title><description>As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.</description><subject>Accuracy</subject><subject>Acoustics</subject><subject>Animals</subject><subject>Classification</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Neural networks</subject><subject>Sound</subject><issn>2071-1050</issn><issn>2071-1050</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNpNUFFLwzAQDqLgmHvxFxR8E6K5pmnSx63qFKaCnfgYsjQd3WozkxTx35s5Qe_l7r7v7r7jQ-gcyBWlBbn2A-TAOaP5ERqlhAMGwsjxv_oUTbzfkBiUQgH5CK1fjLbrvg2t7RPbJG_GB-P6ZNYpvcWl27d1Mm9Xq8iXquuSql33KgyRSWbKRzIS1VTemN6bJxPwolo-4mkIpv-5GaFP67Zn6KRRnTeT3zxGr3e3y_IeL57nD-V0gTWILOCGZ5TphitGVyLn0ACvoQZW1GmeciNExmmqCeNKUZLSIuJACEvjXi6UFnSMLg53d85-DPF7ubGD66OkpEAyYCCixBhdHqa0s94708ida9-V-5JA5N5L-ecl_QaG2mSC</recordid><startdate>20240901</startdate><enddate>20240901</enddate><creator>Zhou, Xiaotao</creator><creator>Wang, Ning</creator><creator>Hu, Kunrong</creator><creator>Wang, Leiguang</creator><creator>Yu, Chunjiang</creator><creator>Guan, Zhenhua</creator><creator>Hu, Ruiqi</creator><creator>Li, Qiumei</creator><creator>Ye, Longjia</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>4U-</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><orcidid>https://orcid.org/0009-0003-9989-5042</orcidid><orcidid>https://orcid.org/0000-0001-8847-4873</orcidid><orcidid>https://orcid.org/0000-0002-6709-4545</orcidid></search><sort><creationdate>20240901</creationdate><title>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</title><author>Zhou, Xiaotao ; Wang, Ning ; Hu, Kunrong ; Wang, Leiguang ; Yu, Chunjiang ; Guan, Zhenhua ; Hu, Ruiqi ; Li, Qiumei ; Ye, Longjia</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c184t-f7435cf7a53b8671f17d1d159d2627e884732c057aa302399d210052f7468ac83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Acoustics</topic><topic>Animals</topic><topic>Classification</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Neural networks</topic><topic>Sound</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Xiaotao</creatorcontrib><creatorcontrib>Wang, Ning</creatorcontrib><creatorcontrib>Hu, Kunrong</creatorcontrib><creatorcontrib>Wang, Leiguang</creatorcontrib><creatorcontrib>Yu, Chunjiang</creatorcontrib><creatorcontrib>Guan, Zhenhua</creatorcontrib><creatorcontrib>Hu, Ruiqi</creatorcontrib><creatorcontrib>Li, Qiumei</creatorcontrib><creatorcontrib>Ye, Longjia</creatorcontrib><collection>CrossRef</collection><collection>University Readers</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><jtitle>Sustainability</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhou, Xiaotao</au><au>Wang, Ning</au><au>Hu, Kunrong</au><au>Wang, Leiguang</au><au>Yu, Chunjiang</au><au>Guan, Zhenhua</au><au>Hu, Ruiqi</au><au>Li, Qiumei</au><au>Ye, Longjia</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network</atitle><jtitle>Sustainability</jtitle><date>2024-09-01</date><risdate>2024</risdate><volume>16</volume><issue>17</issue><spage>7536</spage><pages>7536-</pages><issn>2071-1050</issn><eissn>2071-1050</eissn><abstract>As part of the ecosystem, the western black-crested gibbon (Nomascus concolor) is important for ecological sustainability. Calls are an important means of communication for gibbons, so accurately recognizing and categorizing gibbon calls is important for their population monitoring and conservation. Since a large amount of sound data will be generated in the process of acoustic monitoring, it will take a lot of time to recognize the gibbon calls manually, so this paper proposes a western black-crested gibbon call recognition network based on SA_DenseNet-LSTM-Attention. First, to address the lack of datasets, this paper explores 10 different data extension methods to process all the datasets, and then converts all the sound data into Mel spectrograms for model input. After the test, it is concluded that WaveGAN audio data augmentation method obtains the highest accuracy in improving the classification accuracy of all models in the paper. Then, the method of fusion of DenseNet-extracted features and LSTM-extracted temporal features using PCA principal component analysis is proposed to address the problem of the low accuracy of call recognition, and finally, the SA_DenseNet-LSTM-Attention western black-crested gibbon call recognition network proposed in this paper is used for recognition training. In order to verify the effectiveness of the feature fusion method proposed in this paper, we classified 13 different types of sounds and compared several different networks, and finally, the accuracy of the VGG16 model improved by 2.0%, the accuracy of the Xception model improved by 1.8%, the accuracy of the MobileNet model improved by 2.5%, and the accuracy of the DenseNet network model improved by 2.3%. Compared to other classical chirp recognition networks, our proposed network obtained the highest accuracy of 98.2%, and the convergence of our model is better than all the compared models. Our experiments have demonstrated that the deep learning-based call recognition method can provide better technical support for monitoring western black-crested gibbon populations.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/su16177536</doi><orcidid>https://orcid.org/0009-0003-9989-5042</orcidid><orcidid>https://orcid.org/0000-0001-8847-4873</orcidid><orcidid>https://orcid.org/0000-0002-6709-4545</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2071-1050 |
ispartof | Sustainability, 2024-09, Vol.16 (17), p.7536 |
issn | 2071-1050 2071-1050 |
language | eng |
recordid | cdi_proquest_journals_3104151874 |
source | MDPI - Multidisciplinary Digital Publishing Institute; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals |
subjects | Accuracy Acoustics Animals Classification Datasets Deep learning Neural networks Sound |
title | Recognition of Western Black-Crested Gibbon Call Signatures Based on SA_DenseNet-LSTM-Attention Network |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-03T17%3A00%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Recognition%20of%20Western%20Black-Crested%20Gibbon%20Call%20Signatures%20Based%20on%20SA_DenseNet-LSTM-Attention%20Network&rft.jtitle=Sustainability&rft.au=Zhou,%20Xiaotao&rft.date=2024-09-01&rft.volume=16&rft.issue=17&rft.spage=7536&rft.pages=7536-&rft.issn=2071-1050&rft.eissn=2071-1050&rft_id=info:doi/10.3390/su16177536&rft_dat=%3Cproquest_cross%3E3104151874%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3104151874&rft_id=info:pmid/&rfr_iscdi=true |