Unveiling hidden factors: explainable AI for feature boosting in speech emotion recognition
Speech emotion recognition (SER) has gained significant attention due to its several application fields, such as mental health, education, and human-computer interaction. However, the accuracy of SER systems is hindered by high-dimensional feature sets that may contain irrelevant and redundant infor...
Gespeichert in:
Veröffentlicht in: | Applied intelligence (Dordrecht, Netherlands) Netherlands), 2024-06, Vol.54 (11-12), p.7046-7069 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 7069 |
---|---|
container_issue | 11-12 |
container_start_page | 7046 |
container_title | Applied intelligence (Dordrecht, Netherlands) |
container_volume | 54 |
creator | Nfissi, Alaa Bouachir, Wassim Bouguila, Nizar Mishara, Brian |
description | Speech emotion recognition (SER) has gained significant attention due to its several application fields, such as mental health, education, and human-computer interaction. However, the accuracy of SER systems is hindered by high-dimensional feature sets that may contain irrelevant and redundant information. To overcome this challenge, this study proposes an iterative feature boosting approach for SER that emphasizes feature relevance and explainability to enhance machine learning model performance. Our approach involves meticulous feature selection and analysis to build efficient SER systems. In addressing our main problem through model explainability, we employ a feature evaluation loop with Shapley values to iteratively refine feature sets. This process strikes a balance between model performance and transparency, which enables a comprehensive understanding of the model’s predictions. The proposed approach offers several advantages, including the identification and removal of irrelevant and redundant features, leading to a more effective model. Additionally, it promotes explainability, facilitating comprehension of the model’s predictions and the identification of crucial features for emotion determination. The effectiveness of the proposed method is validated on the SER benchmarks Toronto emotional speech set (TESS), Berlin Database of Emotional Speech (EMO-DB), Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS), and Surrey Audio-Visual Expressed Emotion (SAVEE) dataset, outperforming state-of-the-art methods. These results highlight the potential of the proposed technique in developing accurate and explainable SER systems. To the best of our knowledge, this is the first work to incorporate model explainability into an SER framework. The source code of this paper is publicly available via this
https://github.com/alaaNfissi/Unveiling-Hidden-Factors-Explainable-AI-for-Feature-Boosting-in-Speech-Emotion-Recognition
. |
doi_str_mv | 10.1007/s10489-024-05536-5 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3071639397</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3071639397</sourcerecordid><originalsourceid>FETCH-LOGICAL-c200t-646cad51b0111aa42ee5356a5d107d6edccda0f5f069cf73ca5ea55b29bbcc023</originalsourceid><addsrcrecordid>eNp9kE1LAzEQhoMoWKt_wFPA8-pks8kab6X4USh4sSB4CNnsbJuyTdZkK_rv3bqCN08zMO_zDjyEXDK4ZgDlTWJQ3KoM8iIDIbjMxBGZMFHyrCxUeUwmoIaTlOr1lJyltAUAzoFNyNvKf6BrnV_Tjatr9LQxtg8x3VH87FrjvKlapLMFbUKkDZp-H5FWIaT-wDhPU4doNxR3oXfB04g2rL077OfkpDFtwovfOSWrh_uX-VO2fH5czGfLzOYAfSYLaU0tWAWMMWOKHFFwIY2oGZS1xNra2kAjGpDKNiW3RqARospVVVkLOZ-Sq7G3i-F9j6nX27CPfnipOZRMcsVVOaTyMWVjSClio7vodiZ-aQb6IFGPEvUgUf9I1GKA-AilIezXGP-q_6G-AQQCdmk</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3071639397</pqid></control><display><type>article</type><title>Unveiling hidden factors: explainable AI for feature boosting in speech emotion recognition</title><source>Springer Nature - Complete Springer Journals</source><creator>Nfissi, Alaa ; Bouachir, Wassim ; Bouguila, Nizar ; Mishara, Brian</creator><creatorcontrib>Nfissi, Alaa ; Bouachir, Wassim ; Bouguila, Nizar ; Mishara, Brian</creatorcontrib><description>Speech emotion recognition (SER) has gained significant attention due to its several application fields, such as mental health, education, and human-computer interaction. However, the accuracy of SER systems is hindered by high-dimensional feature sets that may contain irrelevant and redundant information. To overcome this challenge, this study proposes an iterative feature boosting approach for SER that emphasizes feature relevance and explainability to enhance machine learning model performance. Our approach involves meticulous feature selection and analysis to build efficient SER systems. In addressing our main problem through model explainability, we employ a feature evaluation loop with Shapley values to iteratively refine feature sets. This process strikes a balance between model performance and transparency, which enables a comprehensive understanding of the model’s predictions. The proposed approach offers several advantages, including the identification and removal of irrelevant and redundant features, leading to a more effective model. Additionally, it promotes explainability, facilitating comprehension of the model’s predictions and the identification of crucial features for emotion determination. The effectiveness of the proposed method is validated on the SER benchmarks Toronto emotional speech set (TESS), Berlin Database of Emotional Speech (EMO-DB), Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS), and Surrey Audio-Visual Expressed Emotion (SAVEE) dataset, outperforming state-of-the-art methods. These results highlight the potential of the proposed technique in developing accurate and explainable SER systems. To the best of our knowledge, this is the first work to incorporate model explainability into an SER framework. The source code of this paper is publicly available via this
https://github.com/alaaNfissi/Unveiling-Hidden-Factors-Explainable-AI-for-Feature-Boosting-in-Speech-Emotion-Recognition
.</description><identifier>ISSN: 0924-669X</identifier><identifier>EISSN: 1573-7497</identifier><identifier>DOI: 10.1007/s10489-024-05536-5</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial Intelligence ; Computer Science ; Emotion recognition ; Emotions ; Explainable artificial intelligence ; Feature recognition ; Machine learning ; Machines ; Manufacturing ; Mechanical Engineering ; Processes ; Source code ; Speech ; Speech recognition</subject><ispartof>Applied intelligence (Dordrecht, Netherlands), 2024-06, Vol.54 (11-12), p.7046-7069</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c200t-646cad51b0111aa42ee5356a5d107d6edccda0f5f069cf73ca5ea55b29bbcc023</cites><orcidid>0000-0001-7224-7940 ; 0000-0003-3896-7674 ; 0000-0003-2828-8010 ; 0000-0003-2449-6269</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10489-024-05536-5$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10489-024-05536-5$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27903,27904,41467,42536,51298</link.rule.ids></links><search><creatorcontrib>Nfissi, Alaa</creatorcontrib><creatorcontrib>Bouachir, Wassim</creatorcontrib><creatorcontrib>Bouguila, Nizar</creatorcontrib><creatorcontrib>Mishara, Brian</creatorcontrib><title>Unveiling hidden factors: explainable AI for feature boosting in speech emotion recognition</title><title>Applied intelligence (Dordrecht, Netherlands)</title><addtitle>Appl Intell</addtitle><description>Speech emotion recognition (SER) has gained significant attention due to its several application fields, such as mental health, education, and human-computer interaction. However, the accuracy of SER systems is hindered by high-dimensional feature sets that may contain irrelevant and redundant information. To overcome this challenge, this study proposes an iterative feature boosting approach for SER that emphasizes feature relevance and explainability to enhance machine learning model performance. Our approach involves meticulous feature selection and analysis to build efficient SER systems. In addressing our main problem through model explainability, we employ a feature evaluation loop with Shapley values to iteratively refine feature sets. This process strikes a balance between model performance and transparency, which enables a comprehensive understanding of the model’s predictions. The proposed approach offers several advantages, including the identification and removal of irrelevant and redundant features, leading to a more effective model. Additionally, it promotes explainability, facilitating comprehension of the model’s predictions and the identification of crucial features for emotion determination. The effectiveness of the proposed method is validated on the SER benchmarks Toronto emotional speech set (TESS), Berlin Database of Emotional Speech (EMO-DB), Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS), and Surrey Audio-Visual Expressed Emotion (SAVEE) dataset, outperforming state-of-the-art methods. These results highlight the potential of the proposed technique in developing accurate and explainable SER systems. To the best of our knowledge, this is the first work to incorporate model explainability into an SER framework. The source code of this paper is publicly available via this
https://github.com/alaaNfissi/Unveiling-Hidden-Factors-Explainable-AI-for-Feature-Boosting-in-Speech-Emotion-Recognition
.</description><subject>Artificial Intelligence</subject><subject>Computer Science</subject><subject>Emotion recognition</subject><subject>Emotions</subject><subject>Explainable artificial intelligence</subject><subject>Feature recognition</subject><subject>Machine learning</subject><subject>Machines</subject><subject>Manufacturing</subject><subject>Mechanical Engineering</subject><subject>Processes</subject><subject>Source code</subject><subject>Speech</subject><subject>Speech recognition</subject><issn>0924-669X</issn><issn>1573-7497</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kE1LAzEQhoMoWKt_wFPA8-pks8kab6X4USh4sSB4CNnsbJuyTdZkK_rv3bqCN08zMO_zDjyEXDK4ZgDlTWJQ3KoM8iIDIbjMxBGZMFHyrCxUeUwmoIaTlOr1lJyltAUAzoFNyNvKf6BrnV_Tjatr9LQxtg8x3VH87FrjvKlapLMFbUKkDZp-H5FWIaT-wDhPU4doNxR3oXfB04g2rL077OfkpDFtwovfOSWrh_uX-VO2fH5czGfLzOYAfSYLaU0tWAWMMWOKHFFwIY2oGZS1xNra2kAjGpDKNiW3RqARospVVVkLOZ-Sq7G3i-F9j6nX27CPfnipOZRMcsVVOaTyMWVjSClio7vodiZ-aQb6IFGPEvUgUf9I1GKA-AilIezXGP-q_6G-AQQCdmk</recordid><startdate>20240601</startdate><enddate>20240601</enddate><creator>Nfissi, Alaa</creator><creator>Bouachir, Wassim</creator><creator>Bouguila, Nizar</creator><creator>Mishara, Brian</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-7224-7940</orcidid><orcidid>https://orcid.org/0000-0003-3896-7674</orcidid><orcidid>https://orcid.org/0000-0003-2828-8010</orcidid><orcidid>https://orcid.org/0000-0003-2449-6269</orcidid></search><sort><creationdate>20240601</creationdate><title>Unveiling hidden factors: explainable AI for feature boosting in speech emotion recognition</title><author>Nfissi, Alaa ; Bouachir, Wassim ; Bouguila, Nizar ; Mishara, Brian</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c200t-646cad51b0111aa42ee5356a5d107d6edccda0f5f069cf73ca5ea55b29bbcc023</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial Intelligence</topic><topic>Computer Science</topic><topic>Emotion recognition</topic><topic>Emotions</topic><topic>Explainable artificial intelligence</topic><topic>Feature recognition</topic><topic>Machine learning</topic><topic>Machines</topic><topic>Manufacturing</topic><topic>Mechanical Engineering</topic><topic>Processes</topic><topic>Source code</topic><topic>Speech</topic><topic>Speech recognition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Nfissi, Alaa</creatorcontrib><creatorcontrib>Bouachir, Wassim</creatorcontrib><creatorcontrib>Bouguila, Nizar</creatorcontrib><creatorcontrib>Mishara, Brian</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Applied intelligence (Dordrecht, Netherlands)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nfissi, Alaa</au><au>Bouachir, Wassim</au><au>Bouguila, Nizar</au><au>Mishara, Brian</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Unveiling hidden factors: explainable AI for feature boosting in speech emotion recognition</atitle><jtitle>Applied intelligence (Dordrecht, Netherlands)</jtitle><stitle>Appl Intell</stitle><date>2024-06-01</date><risdate>2024</risdate><volume>54</volume><issue>11-12</issue><spage>7046</spage><epage>7069</epage><pages>7046-7069</pages><issn>0924-669X</issn><eissn>1573-7497</eissn><abstract>Speech emotion recognition (SER) has gained significant attention due to its several application fields, such as mental health, education, and human-computer interaction. However, the accuracy of SER systems is hindered by high-dimensional feature sets that may contain irrelevant and redundant information. To overcome this challenge, this study proposes an iterative feature boosting approach for SER that emphasizes feature relevance and explainability to enhance machine learning model performance. Our approach involves meticulous feature selection and analysis to build efficient SER systems. In addressing our main problem through model explainability, we employ a feature evaluation loop with Shapley values to iteratively refine feature sets. This process strikes a balance between model performance and transparency, which enables a comprehensive understanding of the model’s predictions. The proposed approach offers several advantages, including the identification and removal of irrelevant and redundant features, leading to a more effective model. Additionally, it promotes explainability, facilitating comprehension of the model’s predictions and the identification of crucial features for emotion determination. The effectiveness of the proposed method is validated on the SER benchmarks Toronto emotional speech set (TESS), Berlin Database of Emotional Speech (EMO-DB), Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS), and Surrey Audio-Visual Expressed Emotion (SAVEE) dataset, outperforming state-of-the-art methods. These results highlight the potential of the proposed technique in developing accurate and explainable SER systems. To the best of our knowledge, this is the first work to incorporate model explainability into an SER framework. The source code of this paper is publicly available via this
https://github.com/alaaNfissi/Unveiling-Hidden-Factors-Explainable-AI-for-Feature-Boosting-in-Speech-Emotion-Recognition
.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10489-024-05536-5</doi><tpages>24</tpages><orcidid>https://orcid.org/0000-0001-7224-7940</orcidid><orcidid>https://orcid.org/0000-0003-3896-7674</orcidid><orcidid>https://orcid.org/0000-0003-2828-8010</orcidid><orcidid>https://orcid.org/0000-0003-2449-6269</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0924-669X |
ispartof | Applied intelligence (Dordrecht, Netherlands), 2024-06, Vol.54 (11-12), p.7046-7069 |
issn | 0924-669X 1573-7497 |
language | eng |
recordid | cdi_proquest_journals_3071639397 |
source | Springer Nature - Complete Springer Journals |
subjects | Artificial Intelligence Computer Science Emotion recognition Emotions Explainable artificial intelligence Feature recognition Machine learning Machines Manufacturing Mechanical Engineering Processes Source code Speech Speech recognition |
title | Unveiling hidden factors: explainable AI for feature boosting in speech emotion recognition |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-21T14%3A37%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Unveiling%20hidden%20factors:%20explainable%20AI%20for%20feature%20boosting%20in%20speech%20emotion%20recognition&rft.jtitle=Applied%20intelligence%20(Dordrecht,%20Netherlands)&rft.au=Nfissi,%20Alaa&rft.date=2024-06-01&rft.volume=54&rft.issue=11-12&rft.spage=7046&rft.epage=7069&rft.pages=7046-7069&rft.issn=0924-669X&rft.eissn=1573-7497&rft_id=info:doi/10.1007/s10489-024-05536-5&rft_dat=%3Cproquest_cross%3E3071639397%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3071639397&rft_id=info:pmid/&rfr_iscdi=true |