Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection
Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into considerati...
Gespeichert in:
Veröffentlicht in: | IEEE access 2022, Vol.10, p.24976-24984 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 24984 |
---|---|
container_issue | |
container_start_page | 24976 |
container_title | IEEE access |
container_volume | 10 |
creator | Teye, Martha T. Missah, Yaw Marfo Ahene, Emmanuel Frimpong, Twum |
description | Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs. |
doi_str_mv | 10.1109/ACCESS.2022.3153787 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_ACCESS_2022_3153787</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9718284</ieee_id><doaj_id>oai_doaj_org_article_8de9b3c6b2c14361ac4213b7369631d7</doaj_id><sourcerecordid>2638122427</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</originalsourceid><addsrcrecordid>eNpNUcFu3CAQRVUrNdrkC3JB6rW7BQYD7m3luk2kSD2kOSOM8dYrL6SAV83fF6-jqHOAmad5bwYeQreU7Cgl9Zd907SPjztGGNsBrUAq-Q5dMSrqLVQg3v-Xf0Q3KR1JCVWgSl6h3-3ZTLPJY_A4DLgJ_uxiutRmwvuD8zl9xU--L2g2vh_9ATfzlOfoPi_d2f3NuOC49ecxBn8qBDx63J7CRfOby84u2TX6MJgpuZvXe4Oevre_mrvtw88f983-YWs5UXlLoZeDpaIqTysxmI4z7ng5JfSdIBIcVUKCrZSjIGpieKeAEAtcABl62KD7VbcP5qif43gy8UUHM-oLEOJBm5hHOzmteld3YEXHLOUgqLGcUehkkRVAe1m0Pq1azzH8mV3K-hjmWD4maSZAUcaWtTYI1i4bQ0rRDW9TKdGLQ3p1SC8O6VeHCut2ZY3OuTdGLaliisM_pnKLhQ</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2638122427</pqid></control><display><type>article</type><title>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Teye, Martha T. ; Missah, Yaw Marfo ; Ahene, Emmanuel ; Frimpong, Twum</creator><creatorcontrib>Teye, Martha T. ; Missah, Yaw Marfo ; Ahene, Emmanuel ; Frimpong, Twum</creatorcontrib><description>Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2022.3153787</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Agents (artificial intelligence) ; AI ethics ; Algorithms ; Artificial neural networks ; Biological system modeling ; Biometrics ; Conversational artificial intelligence ; convolutional neural network ; Convolutional neural networks ; Cultural factors ; Data models ; Decision analysis ; Emotion recognition ; Emotions ; Face recognition ; Geographical locations ; human-AI interaction ; Model accuracy ; Prediction models ; Social networking (online) ; Speech recognition</subject><ispartof>IEEE access, 2022, Vol.10, p.24976-24984</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</citedby><cites>FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</cites><orcidid>0000-0002-2370-4700 ; 0000-0002-2926-681X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9718284$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,777,781,861,2096,4010,27614,27904,27905,27906,54914</link.rule.ids></links><search><creatorcontrib>Teye, Martha T.</creatorcontrib><creatorcontrib>Missah, Yaw Marfo</creatorcontrib><creatorcontrib>Ahene, Emmanuel</creatorcontrib><creatorcontrib>Frimpong, Twum</creatorcontrib><title>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</title><title>IEEE access</title><addtitle>Access</addtitle><description>Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.</description><subject>Agents (artificial intelligence)</subject><subject>AI ethics</subject><subject>Algorithms</subject><subject>Artificial neural networks</subject><subject>Biological system modeling</subject><subject>Biometrics</subject><subject>Conversational artificial intelligence</subject><subject>convolutional neural network</subject><subject>Convolutional neural networks</subject><subject>Cultural factors</subject><subject>Data models</subject><subject>Decision analysis</subject><subject>Emotion recognition</subject><subject>Emotions</subject><subject>Face recognition</subject><subject>Geographical locations</subject><subject>human-AI interaction</subject><subject>Model accuracy</subject><subject>Prediction models</subject><subject>Social networking (online)</subject><subject>Speech recognition</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUcFu3CAQRVUrNdrkC3JB6rW7BQYD7m3luk2kSD2kOSOM8dYrL6SAV83fF6-jqHOAmad5bwYeQreU7Cgl9Zd907SPjztGGNsBrUAq-Q5dMSrqLVQg3v-Xf0Q3KR1JCVWgSl6h3-3ZTLPJY_A4DLgJ_uxiutRmwvuD8zl9xU--L2g2vh_9ATfzlOfoPi_d2f3NuOC49ecxBn8qBDx63J7CRfOby84u2TX6MJgpuZvXe4Oevre_mrvtw88f983-YWs5UXlLoZeDpaIqTysxmI4z7ng5JfSdIBIcVUKCrZSjIGpieKeAEAtcABl62KD7VbcP5qif43gy8UUHM-oLEOJBm5hHOzmteld3YEXHLOUgqLGcUehkkRVAe1m0Pq1azzH8mV3K-hjmWD4maSZAUcaWtTYI1i4bQ0rRDW9TKdGLQ3p1SC8O6VeHCut2ZY3OuTdGLaliisM_pnKLhQ</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Teye, Martha T.</creator><creator>Missah, Yaw Marfo</creator><creator>Ahene, Emmanuel</creator><creator>Frimpong, Twum</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-2370-4700</orcidid><orcidid>https://orcid.org/0000-0002-2926-681X</orcidid></search><sort><creationdate>2022</creationdate><title>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</title><author>Teye, Martha T. ; Missah, Yaw Marfo ; Ahene, Emmanuel ; Frimpong, Twum</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Agents (artificial intelligence)</topic><topic>AI ethics</topic><topic>Algorithms</topic><topic>Artificial neural networks</topic><topic>Biological system modeling</topic><topic>Biometrics</topic><topic>Conversational artificial intelligence</topic><topic>convolutional neural network</topic><topic>Convolutional neural networks</topic><topic>Cultural factors</topic><topic>Data models</topic><topic>Decision analysis</topic><topic>Emotion recognition</topic><topic>Emotions</topic><topic>Face recognition</topic><topic>Geographical locations</topic><topic>human-AI interaction</topic><topic>Model accuracy</topic><topic>Prediction models</topic><topic>Social networking (online)</topic><topic>Speech recognition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Teye, Martha T.</creatorcontrib><creatorcontrib>Missah, Yaw Marfo</creatorcontrib><creatorcontrib>Ahene, Emmanuel</creatorcontrib><creatorcontrib>Frimpong, Twum</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Teye, Martha T.</au><au>Missah, Yaw Marfo</au><au>Ahene, Emmanuel</au><au>Frimpong, Twum</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2022</date><risdate>2022</risdate><volume>10</volume><spage>24976</spage><epage>24984</epage><pages>24976-24984</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2022.3153787</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0002-2370-4700</orcidid><orcidid>https://orcid.org/0000-0002-2926-681X</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2022, Vol.10, p.24976-24984 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_crossref_primary_10_1109_ACCESS_2022_3153787 |
source | IEEE Open Access Journals; DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals |
subjects | Agents (artificial intelligence) AI ethics Algorithms Artificial neural networks Biological system modeling Biometrics Conversational artificial intelligence convolutional neural network Convolutional neural networks Cultural factors Data models Decision analysis Emotion recognition Emotions Face recognition Geographical locations human-AI interaction Model accuracy Prediction models Social networking (online) Speech recognition |
title | Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-19T06%3A26%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluation%20of%20Conversational%20Agents:%20Understanding%20Culture,%20Context%20and%20Environment%20in%20Emotion%20Detection&rft.jtitle=IEEE%20access&rft.au=Teye,%20Martha%20T.&rft.date=2022&rft.volume=10&rft.spage=24976&rft.epage=24984&rft.pages=24976-24984&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2022.3153787&rft_dat=%3Cproquest_cross%3E2638122427%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2638122427&rft_id=info:pmid/&rft_ieee_id=9718284&rft_doaj_id=oai_doaj_org_article_8de9b3c6b2c14361ac4213b7369631d7&rfr_iscdi=true |