Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection

Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into considerati...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2022, Vol.10, p.24976-24984
Hauptverfasser: Teye, Martha T., Missah, Yaw Marfo, Ahene, Emmanuel, Frimpong, Twum
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 24984
container_issue
container_start_page 24976
container_title IEEE access
container_volume 10
creator Teye, Martha T.
Missah, Yaw Marfo
Ahene, Emmanuel
Frimpong, Twum
description Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.
doi_str_mv 10.1109/ACCESS.2022.3153787
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_ACCESS_2022_3153787</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9718284</ieee_id><doaj_id>oai_doaj_org_article_8de9b3c6b2c14361ac4213b7369631d7</doaj_id><sourcerecordid>2638122427</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</originalsourceid><addsrcrecordid>eNpNUcFu3CAQRVUrNdrkC3JB6rW7BQYD7m3luk2kSD2kOSOM8dYrL6SAV83fF6-jqHOAmad5bwYeQreU7Cgl9Zd907SPjztGGNsBrUAq-Q5dMSrqLVQg3v-Xf0Q3KR1JCVWgSl6h3-3ZTLPJY_A4DLgJ_uxiutRmwvuD8zl9xU--L2g2vh_9ATfzlOfoPi_d2f3NuOC49ecxBn8qBDx63J7CRfOby84u2TX6MJgpuZvXe4Oevre_mrvtw88f983-YWs5UXlLoZeDpaIqTysxmI4z7ng5JfSdIBIcVUKCrZSjIGpieKeAEAtcABl62KD7VbcP5qif43gy8UUHM-oLEOJBm5hHOzmteld3YEXHLOUgqLGcUehkkRVAe1m0Pq1azzH8mV3K-hjmWD4maSZAUcaWtTYI1i4bQ0rRDW9TKdGLQ3p1SC8O6VeHCut2ZY3OuTdGLaliisM_pnKLhQ</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2638122427</pqid></control><display><type>article</type><title>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Teye, Martha T. ; Missah, Yaw Marfo ; Ahene, Emmanuel ; Frimpong, Twum</creator><creatorcontrib>Teye, Martha T. ; Missah, Yaw Marfo ; Ahene, Emmanuel ; Frimpong, Twum</creatorcontrib><description>Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2022.3153787</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Agents (artificial intelligence) ; AI ethics ; Algorithms ; Artificial neural networks ; Biological system modeling ; Biometrics ; Conversational artificial intelligence ; convolutional neural network ; Convolutional neural networks ; Cultural factors ; Data models ; Decision analysis ; Emotion recognition ; Emotions ; Face recognition ; Geographical locations ; human-AI interaction ; Model accuracy ; Prediction models ; Social networking (online) ; Speech recognition</subject><ispartof>IEEE access, 2022, Vol.10, p.24976-24984</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</citedby><cites>FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</cites><orcidid>0000-0002-2370-4700 ; 0000-0002-2926-681X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9718284$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,777,781,861,2096,4010,27614,27904,27905,27906,54914</link.rule.ids></links><search><creatorcontrib>Teye, Martha T.</creatorcontrib><creatorcontrib>Missah, Yaw Marfo</creatorcontrib><creatorcontrib>Ahene, Emmanuel</creatorcontrib><creatorcontrib>Frimpong, Twum</creatorcontrib><title>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</title><title>IEEE access</title><addtitle>Access</addtitle><description>Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.</description><subject>Agents (artificial intelligence)</subject><subject>AI ethics</subject><subject>Algorithms</subject><subject>Artificial neural networks</subject><subject>Biological system modeling</subject><subject>Biometrics</subject><subject>Conversational artificial intelligence</subject><subject>convolutional neural network</subject><subject>Convolutional neural networks</subject><subject>Cultural factors</subject><subject>Data models</subject><subject>Decision analysis</subject><subject>Emotion recognition</subject><subject>Emotions</subject><subject>Face recognition</subject><subject>Geographical locations</subject><subject>human-AI interaction</subject><subject>Model accuracy</subject><subject>Prediction models</subject><subject>Social networking (online)</subject><subject>Speech recognition</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUcFu3CAQRVUrNdrkC3JB6rW7BQYD7m3luk2kSD2kOSOM8dYrL6SAV83fF6-jqHOAmad5bwYeQreU7Cgl9Zd907SPjztGGNsBrUAq-Q5dMSrqLVQg3v-Xf0Q3KR1JCVWgSl6h3-3ZTLPJY_A4DLgJ_uxiutRmwvuD8zl9xU--L2g2vh_9ATfzlOfoPi_d2f3NuOC49ecxBn8qBDx63J7CRfOby84u2TX6MJgpuZvXe4Oevre_mrvtw88f983-YWs5UXlLoZeDpaIqTysxmI4z7ng5JfSdIBIcVUKCrZSjIGpieKeAEAtcABl62KD7VbcP5qif43gy8UUHM-oLEOJBm5hHOzmteld3YEXHLOUgqLGcUehkkRVAe1m0Pq1azzH8mV3K-hjmWD4maSZAUcaWtTYI1i4bQ0rRDW9TKdGLQ3p1SC8O6VeHCut2ZY3OuTdGLaliisM_pnKLhQ</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Teye, Martha T.</creator><creator>Missah, Yaw Marfo</creator><creator>Ahene, Emmanuel</creator><creator>Frimpong, Twum</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-2370-4700</orcidid><orcidid>https://orcid.org/0000-0002-2926-681X</orcidid></search><sort><creationdate>2022</creationdate><title>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</title><author>Teye, Martha T. ; Missah, Yaw Marfo ; Ahene, Emmanuel ; Frimpong, Twum</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-13d7fc165110000fab424e4b4273db6073e18673c58e13690a4b8300c34630fd3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Agents (artificial intelligence)</topic><topic>AI ethics</topic><topic>Algorithms</topic><topic>Artificial neural networks</topic><topic>Biological system modeling</topic><topic>Biometrics</topic><topic>Conversational artificial intelligence</topic><topic>convolutional neural network</topic><topic>Convolutional neural networks</topic><topic>Cultural factors</topic><topic>Data models</topic><topic>Decision analysis</topic><topic>Emotion recognition</topic><topic>Emotions</topic><topic>Face recognition</topic><topic>Geographical locations</topic><topic>human-AI interaction</topic><topic>Model accuracy</topic><topic>Prediction models</topic><topic>Social networking (online)</topic><topic>Speech recognition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Teye, Martha T.</creatorcontrib><creatorcontrib>Missah, Yaw Marfo</creatorcontrib><creatorcontrib>Ahene, Emmanuel</creatorcontrib><creatorcontrib>Frimpong, Twum</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Teye, Martha T.</au><au>Missah, Yaw Marfo</au><au>Ahene, Emmanuel</au><au>Frimpong, Twum</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2022</date><risdate>2022</risdate><volume>10</volume><spage>24976</spage><epage>24984</epage><pages>24976-24984</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Valuable decisions and highly prioritized analysis now depend on applications such as facial biometrics, social media photo tagging, and human robots interactions. However, the ability to successfully deploy such applications is based on their efficiencies on tested use cases taking into consideration possible edge cases. Over the years, lots of generalized solutions have been implemented to mimic human emotions including sarcasm. However, factors such as geographical location or cultural difference have not been explored fully amidst its relevance in resolving ethical issues and improving conversational AI (Artificial Intelligence). In this paper, we seek to address the potential challenges in the usage of conversational AI within Black African society. We develop an emotion prediction model with accuracies ranging between 85% and 96%. Our model combines both speech and image data to detect the seven basic emotions with a focus on also identifying sarcasm. It uses 3-layers of the Convolutional Neural Network in addition to a new Audio-Frame Mean Expression (AFME) algorithm and focuses on model pre-processing and post-processing stages. In the end, our proposed solution contributes to maintaining the credibility of an emotion recognition system in conversational AIs.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2022.3153787</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0002-2370-4700</orcidid><orcidid>https://orcid.org/0000-0002-2926-681X</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2022, Vol.10, p.24976-24984
issn 2169-3536
2169-3536
language eng
recordid cdi_crossref_primary_10_1109_ACCESS_2022_3153787
source IEEE Open Access Journals; DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals
subjects Agents (artificial intelligence)
AI ethics
Algorithms
Artificial neural networks
Biological system modeling
Biometrics
Conversational artificial intelligence
convolutional neural network
Convolutional neural networks
Cultural factors
Data models
Decision analysis
Emotion recognition
Emotions
Face recognition
Geographical locations
human-AI interaction
Model accuracy
Prediction models
Social networking (online)
Speech recognition
title Evaluation of Conversational Agents: Understanding Culture, Context and Environment in Emotion Detection
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-19T06%3A26%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluation%20of%20Conversational%20Agents:%20Understanding%20Culture,%20Context%20and%20Environment%20in%20Emotion%20Detection&rft.jtitle=IEEE%20access&rft.au=Teye,%20Martha%20T.&rft.date=2022&rft.volume=10&rft.spage=24976&rft.epage=24984&rft.pages=24976-24984&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2022.3153787&rft_dat=%3Cproquest_cross%3E2638122427%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2638122427&rft_id=info:pmid/&rft_ieee_id=9718284&rft_doaj_id=oai_doaj_org_article_8de9b3c6b2c14361ac4213b7369631d7&rfr_iscdi=true