Discussions of different deep transfer learning models for emotion recognitions
In recent years, facial emotion recognition (FER) has been a popular topic in affective computing. However, FER still faces many challenges in automatic recognition for several reasons, including quality control of sample data, extraction of effective features, creation of models, and multi-feature...
Gespeichert in:
Veröffentlicht in: | IEEE access 2022, Vol.10, p.1-1 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1 |
---|---|
container_issue | |
container_start_page | 1 |
container_title | IEEE access |
container_volume | 10 |
creator | Yen, Chih-Ta Li, Kang-Hua |
description | In recent years, facial emotion recognition (FER) has been a popular topic in affective computing. However, FER still faces many challenges in automatic recognition for several reasons, including quality control of sample data, extraction of effective features, creation of models, and multi-feature fusion, which have not been thoroughly researched and therefore are still hot topics in computer visualization. In view of the mature development of deep learning, deep learning methods are increasingly being used in FER. However, because deep learning requires a large amount of data to achieve effective training, many studies have employed transfer learning to compensate for this drawback. Nevertheless, there has been no universal approach for transfer learning in FER. Accordingly, this study used the five classic models in FER (i.e., ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet121) to conduct a series of experiments: data preprocessing, training type, and the applicability of multi-stage pretraining. According to the results, class wight was the optimal technique for data balance. In addition, the freeze + fine-tuning training type can produce higher accuracy, regardless of the size of the dataset. Multi-stage training was also effective. Compared with the model accuracy in previous studies, the accuracy achieved in this study using the proposed transfer learning method was superior for both large and small datasets. Specifically, on AffectNet, the accuracy for the ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet-121 models increased by 8.37%, 10.45%, 10.45%, 8.55%, and 5.47%, respectively. On FER2013, the accuracy for these models increased by 5.72%, 2%, 10.45%, 5%, and 9%, respectively. These results proved the validity and advantages of the experiments in this study. |
doi_str_mv | 10.1109/ACCESS.2022.3209813 |
format | Article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_9903451</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9903451</ieee_id><doaj_id>oai_doaj_org_article_ee90c00d2d5a4d7c9cc16bb934fd4e4f</doaj_id><sourcerecordid>2721429881</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-a7903225afa8eb9741ef9c6de38a4f7e40d652a36a619a5828126066d41317763</originalsourceid><addsrcrecordid>eNpNUU1rAjEQDaWFivUXeAn0rM3XZjdHsbYVBA-25xCTiUR0Y5P10H_f2BXpXGZ4zHvz8RAaUzKllKiX2Xy-2GymjDA25YyohvI7NGBUqgmvuLz_Vz-iUc57UqIpUFUP0Po1ZHvOOcQ24-ixC95DgrbDDuCEu2TaXAB8AJPa0O7wMTo4ZOxjwnCMXeHhBDbu2nCp8xN68OaQYXTNQ_T1tvicf0xW6_flfLaaWEGabmJqRThjlfGmga2qBQWvrHTAGyN8DYI4WTHDpZFUmaphDWWSSOkE5bSuJR-iZa_rotnrUwpHk350NEH_ATHttEldsAfQAIpYQhxzlRGutspaKrdbxYV3AoQvWs-91inF7zPkTu_jObVlfc1qRgVTTfnpEPG-y6aYcwJ_m0qJvhiheyP0xQh9NaKwxj0rAMCNocr1oqL8F6d9hDk</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2721429881</pqid></control><display><type>article</type><title>Discussions of different deep transfer learning models for emotion recognitions</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Yen, Chih-Ta ; Li, Kang-Hua</creator><creatorcontrib>Yen, Chih-Ta ; Li, Kang-Hua</creatorcontrib><description>In recent years, facial emotion recognition (FER) has been a popular topic in affective computing. However, FER still faces many challenges in automatic recognition for several reasons, including quality control of sample data, extraction of effective features, creation of models, and multi-feature fusion, which have not been thoroughly researched and therefore are still hot topics in computer visualization. In view of the mature development of deep learning, deep learning methods are increasingly being used in FER. However, because deep learning requires a large amount of data to achieve effective training, many studies have employed transfer learning to compensate for this drawback. Nevertheless, there has been no universal approach for transfer learning in FER. Accordingly, this study used the five classic models in FER (i.e., ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet121) to conduct a series of experiments: data preprocessing, training type, and the applicability of multi-stage pretraining. According to the results, class wight was the optimal technique for data balance. In addition, the freeze + fine-tuning training type can produce higher accuracy, regardless of the size of the dataset. Multi-stage training was also effective. Compared with the model accuracy in previous studies, the accuracy achieved in this study using the proposed transfer learning method was superior for both large and small datasets. Specifically, on AffectNet, the accuracy for the ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet-121 models increased by 8.37%, 10.45%, 10.45%, 8.55%, and 5.47%, respectively. On FER2013, the accuracy for these models increased by 5.72%, 2%, 10.45%, 5%, and 9%, respectively. These results proved the validity and advantages of the experiments in this study.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2022.3209813</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Accuracy ; Affective computing ; AffectNet ; convolutional neural network ; Convolutional neural networks ; Data models ; Datasets ; Deep learning ; Emotion recognition ; Emotions ; facial emotion recognition (FER) ; Feature extraction ; FER2013 ; fine-tuning ; Machine learning ; Model accuracy ; pretrained models ; Quality control ; Task analysis ; Training ; Transfer learning</subject><ispartof>IEEE access, 2022, Vol.10, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-a7903225afa8eb9741ef9c6de38a4f7e40d652a36a619a5828126066d41317763</citedby><cites>FETCH-LOGICAL-c408t-a7903225afa8eb9741ef9c6de38a4f7e40d652a36a619a5828126066d41317763</cites><orcidid>0000-0003-0831-6774</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9903451$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,777,781,861,2096,4010,27614,27904,27905,27906,54914</link.rule.ids></links><search><creatorcontrib>Yen, Chih-Ta</creatorcontrib><creatorcontrib>Li, Kang-Hua</creatorcontrib><title>Discussions of different deep transfer learning models for emotion recognitions</title><title>IEEE access</title><addtitle>Access</addtitle><description>In recent years, facial emotion recognition (FER) has been a popular topic in affective computing. However, FER still faces many challenges in automatic recognition for several reasons, including quality control of sample data, extraction of effective features, creation of models, and multi-feature fusion, which have not been thoroughly researched and therefore are still hot topics in computer visualization. In view of the mature development of deep learning, deep learning methods are increasingly being used in FER. However, because deep learning requires a large amount of data to achieve effective training, many studies have employed transfer learning to compensate for this drawback. Nevertheless, there has been no universal approach for transfer learning in FER. Accordingly, this study used the five classic models in FER (i.e., ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet121) to conduct a series of experiments: data preprocessing, training type, and the applicability of multi-stage pretraining. According to the results, class wight was the optimal technique for data balance. In addition, the freeze + fine-tuning training type can produce higher accuracy, regardless of the size of the dataset. Multi-stage training was also effective. Compared with the model accuracy in previous studies, the accuracy achieved in this study using the proposed transfer learning method was superior for both large and small datasets. Specifically, on AffectNet, the accuracy for the ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet-121 models increased by 8.37%, 10.45%, 10.45%, 8.55%, and 5.47%, respectively. On FER2013, the accuracy for these models increased by 5.72%, 2%, 10.45%, 5%, and 9%, respectively. These results proved the validity and advantages of the experiments in this study.</description><subject>Accuracy</subject><subject>Affective computing</subject><subject>AffectNet</subject><subject>convolutional neural network</subject><subject>Convolutional neural networks</subject><subject>Data models</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Emotion recognition</subject><subject>Emotions</subject><subject>facial emotion recognition (FER)</subject><subject>Feature extraction</subject><subject>FER2013</subject><subject>fine-tuning</subject><subject>Machine learning</subject><subject>Model accuracy</subject><subject>pretrained models</subject><subject>Quality control</subject><subject>Task analysis</subject><subject>Training</subject><subject>Transfer learning</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUU1rAjEQDaWFivUXeAn0rM3XZjdHsbYVBA-25xCTiUR0Y5P10H_f2BXpXGZ4zHvz8RAaUzKllKiX2Xy-2GymjDA25YyohvI7NGBUqgmvuLz_Vz-iUc57UqIpUFUP0Po1ZHvOOcQ24-ixC95DgrbDDuCEu2TaXAB8AJPa0O7wMTo4ZOxjwnCMXeHhBDbu2nCp8xN68OaQYXTNQ_T1tvicf0xW6_flfLaaWEGabmJqRThjlfGmga2qBQWvrHTAGyN8DYI4WTHDpZFUmaphDWWSSOkE5bSuJR-iZa_rotnrUwpHk350NEH_ATHttEldsAfQAIpYQhxzlRGutspaKrdbxYV3AoQvWs-91inF7zPkTu_jObVlfc1qRgVTTfnpEPG-y6aYcwJ_m0qJvhiheyP0xQh9NaKwxj0rAMCNocr1oqL8F6d9hDk</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Yen, Chih-Ta</creator><creator>Li, Kang-Hua</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0003-0831-6774</orcidid></search><sort><creationdate>2022</creationdate><title>Discussions of different deep transfer learning models for emotion recognitions</title><author>Yen, Chih-Ta ; Li, Kang-Hua</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-a7903225afa8eb9741ef9c6de38a4f7e40d652a36a619a5828126066d41317763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Accuracy</topic><topic>Affective computing</topic><topic>AffectNet</topic><topic>convolutional neural network</topic><topic>Convolutional neural networks</topic><topic>Data models</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Emotion recognition</topic><topic>Emotions</topic><topic>facial emotion recognition (FER)</topic><topic>Feature extraction</topic><topic>FER2013</topic><topic>fine-tuning</topic><topic>Machine learning</topic><topic>Model accuracy</topic><topic>pretrained models</topic><topic>Quality control</topic><topic>Task analysis</topic><topic>Training</topic><topic>Transfer learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yen, Chih-Ta</creatorcontrib><creatorcontrib>Li, Kang-Hua</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yen, Chih-Ta</au><au>Li, Kang-Hua</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Discussions of different deep transfer learning models for emotion recognitions</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2022</date><risdate>2022</risdate><volume>10</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>In recent years, facial emotion recognition (FER) has been a popular topic in affective computing. However, FER still faces many challenges in automatic recognition for several reasons, including quality control of sample data, extraction of effective features, creation of models, and multi-feature fusion, which have not been thoroughly researched and therefore are still hot topics in computer visualization. In view of the mature development of deep learning, deep learning methods are increasingly being used in FER. However, because deep learning requires a large amount of data to achieve effective training, many studies have employed transfer learning to compensate for this drawback. Nevertheless, there has been no universal approach for transfer learning in FER. Accordingly, this study used the five classic models in FER (i.e., ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet121) to conduct a series of experiments: data preprocessing, training type, and the applicability of multi-stage pretraining. According to the results, class wight was the optimal technique for data balance. In addition, the freeze + fine-tuning training type can produce higher accuracy, regardless of the size of the dataset. Multi-stage training was also effective. Compared with the model accuracy in previous studies, the accuracy achieved in this study using the proposed transfer learning method was superior for both large and small datasets. Specifically, on AffectNet, the accuracy for the ResNet-50, Xception, EfficientNet-B0, Inception, and DenseNet-121 models increased by 8.37%, 10.45%, 10.45%, 8.55%, and 5.47%, respectively. On FER2013, the accuracy for these models increased by 5.72%, 2%, 10.45%, 5%, and 9%, respectively. These results proved the validity and advantages of the experiments in this study.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2022.3209813</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0003-0831-6774</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2022, Vol.10, p.1-1 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_ieee_primary_9903451 |
source | IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals |
subjects | Accuracy Affective computing AffectNet convolutional neural network Convolutional neural networks Data models Datasets Deep learning Emotion recognition Emotions facial emotion recognition (FER) Feature extraction FER2013 fine-tuning Machine learning Model accuracy pretrained models Quality control Task analysis Training Transfer learning |
title | Discussions of different deep transfer learning models for emotion recognitions |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-18T15%3A00%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Discussions%20of%20different%20deep%20transfer%20learning%20models%20for%20emotion%20recognitions&rft.jtitle=IEEE%20access&rft.au=Yen,%20Chih-Ta&rft.date=2022&rft.volume=10&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2022.3209813&rft_dat=%3Cproquest_ieee_%3E2721429881%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2721429881&rft_id=info:pmid/&rft_ieee_id=9903451&rft_doaj_id=oai_doaj_org_article_ee90c00d2d5a4d7c9cc16bb934fd4e4f&rfr_iscdi=true |