Multi-label Learning Based Deep Transfer Neural Network for Facial Attribute Classification
Deep Neural Network (DNN) has recently achieved outstanding performance in a variety of computer vision tasks, including facial attribute classification. The great success of classifying facial attributes with DNN often relies on a massive amount of labelled data. However, in real-world applications...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2018-05 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Zhuang, Ni Yan, Yan Chen, Si Wang, Hanzi Shen, Chunhua |
description | Deep Neural Network (DNN) has recently achieved outstanding performance in a variety of computer vision tasks, including facial attribute classification. The great success of classifying facial attributes with DNN often relies on a massive amount of labelled data. However, in real-world applications, labelled data are only provided for some commonly used attributes (such as age, gender); whereas, unlabelled data are available for other attributes (such as attraction, hairline). To address the above problem, we propose a novel deep transfer neural network method based on multi-label learning for facial attribute classification, termed FMTNet, which consists of three sub-networks: the Face detection Network (FNet), the Multi-label learning Network (MNet) and the Transfer learning Network (TNet). Firstly, based on the Faster Region-based Convolutional Neural Network (Faster R-CNN), FNet is fine-tuned for face detection. Then, MNet is fine-tuned by FNet to predict multiple attributes with labelled data, where an effective loss weight scheme is developed to explicitly exploit the correlation between facial attributes based on attribute grouping. Finally, based on MNet, TNet is trained by taking advantage of unsupervised domain adaptation for unlabelled facial attribute classification. The three sub-networks are tightly coupled to perform effective facial attribute classification. A distinguishing characteristic of the proposed FMTNet method is that the three sub-networks (FNet, MNet and TNet) are constructed in a similar network structure. Extensive experimental results on challenging face datasets demonstrate the effectiveness of our proposed method compared with several state-of-the-art methods. |
doi_str_mv | 10.48550/arxiv.1805.01282 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_1805_01282</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2072249894</sourcerecordid><originalsourceid>FETCH-LOGICAL-a524-a06060bb99e199dbe874a833856741fcec453a9a3ac4411ab831164377ab859f3</originalsourceid><addsrcrecordid>eNotkEtPwzAQhC0kJKrSH8AJS5wT_EzsYym0IBW45MYh2qQb5BKSYDs8_j2hRXuY1ejTamcIueAsVUZrdg3-232m3DCdMi6MOCEzISVPjBLijCxC2DPGRJYLreWMvDyObXRJCxW2dIvgO9e90hsIuKO3iAMtPHShQU-fcPTQThK_ev9Gm97TNdRuspYxeleNEemqhRBc42qIru_OyWkDbcDFv85Jsb4rVvfJ9nnzsFpuE9BCJcCyaarKWuTW7io0uQIjpdFZrnhTY620BAsSaqU4h8pIzjMl83xatW3knFwezx6Sl4N37-B_yr8GykMDE3F1JAbff4wYYrnvR99NP5WC5UIoa6ySv2aMXbY</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2072249894</pqid></control><display><type>article</type><title>Multi-label Learning Based Deep Transfer Neural Network for Facial Attribute Classification</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Zhuang, Ni ; Yan, Yan ; Chen, Si ; Wang, Hanzi ; Shen, Chunhua</creator><creatorcontrib>Zhuang, Ni ; Yan, Yan ; Chen, Si ; Wang, Hanzi ; Shen, Chunhua</creatorcontrib><description>Deep Neural Network (DNN) has recently achieved outstanding performance in a variety of computer vision tasks, including facial attribute classification. The great success of classifying facial attributes with DNN often relies on a massive amount of labelled data. However, in real-world applications, labelled data are only provided for some commonly used attributes (such as age, gender); whereas, unlabelled data are available for other attributes (such as attraction, hairline). To address the above problem, we propose a novel deep transfer neural network method based on multi-label learning for facial attribute classification, termed FMTNet, which consists of three sub-networks: the Face detection Network (FNet), the Multi-label learning Network (MNet) and the Transfer learning Network (TNet). Firstly, based on the Faster Region-based Convolutional Neural Network (Faster R-CNN), FNet is fine-tuned for face detection. Then, MNet is fine-tuned by FNet to predict multiple attributes with labelled data, where an effective loss weight scheme is developed to explicitly exploit the correlation between facial attributes based on attribute grouping. Finally, based on MNet, TNet is trained by taking advantage of unsupervised domain adaptation for unlabelled facial attribute classification. The three sub-networks are tightly coupled to perform effective facial attribute classification. A distinguishing characteristic of the proposed FMTNet method is that the three sub-networks (FNet, MNet and TNet) are constructed in a similar network structure. Extensive experimental results on challenging face datasets demonstrate the effectiveness of our proposed method compared with several state-of-the-art methods.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.1805.01282</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Artificial neural networks ; Classification ; Computer Science - Computer Vision and Pattern Recognition ; Computer vision ; Face recognition ; Learning ; Neural networks ; Weight</subject><ispartof>arXiv.org, 2018-05</ispartof><rights>2018. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,784,885,27924</link.rule.ids><backlink>$$Uhttps://doi.org/10.1016/j.patcog.2018.03.018$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.48550/arXiv.1805.01282$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhuang, Ni</creatorcontrib><creatorcontrib>Yan, Yan</creatorcontrib><creatorcontrib>Chen, Si</creatorcontrib><creatorcontrib>Wang, Hanzi</creatorcontrib><creatorcontrib>Shen, Chunhua</creatorcontrib><title>Multi-label Learning Based Deep Transfer Neural Network for Facial Attribute Classification</title><title>arXiv.org</title><description>Deep Neural Network (DNN) has recently achieved outstanding performance in a variety of computer vision tasks, including facial attribute classification. The great success of classifying facial attributes with DNN often relies on a massive amount of labelled data. However, in real-world applications, labelled data are only provided for some commonly used attributes (such as age, gender); whereas, unlabelled data are available for other attributes (such as attraction, hairline). To address the above problem, we propose a novel deep transfer neural network method based on multi-label learning for facial attribute classification, termed FMTNet, which consists of three sub-networks: the Face detection Network (FNet), the Multi-label learning Network (MNet) and the Transfer learning Network (TNet). Firstly, based on the Faster Region-based Convolutional Neural Network (Faster R-CNN), FNet is fine-tuned for face detection. Then, MNet is fine-tuned by FNet to predict multiple attributes with labelled data, where an effective loss weight scheme is developed to explicitly exploit the correlation between facial attributes based on attribute grouping. Finally, based on MNet, TNet is trained by taking advantage of unsupervised domain adaptation for unlabelled facial attribute classification. The three sub-networks are tightly coupled to perform effective facial attribute classification. A distinguishing characteristic of the proposed FMTNet method is that the three sub-networks (FNet, MNet and TNet) are constructed in a similar network structure. Extensive experimental results on challenging face datasets demonstrate the effectiveness of our proposed method compared with several state-of-the-art methods.</description><subject>Artificial neural networks</subject><subject>Classification</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer vision</subject><subject>Face recognition</subject><subject>Learning</subject><subject>Neural networks</subject><subject>Weight</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GOX</sourceid><recordid>eNotkEtPwzAQhC0kJKrSH8AJS5wT_EzsYym0IBW45MYh2qQb5BKSYDs8_j2hRXuY1ejTamcIueAsVUZrdg3-232m3DCdMi6MOCEzISVPjBLijCxC2DPGRJYLreWMvDyObXRJCxW2dIvgO9e90hsIuKO3iAMtPHShQU-fcPTQThK_ev9Gm97TNdRuspYxeleNEemqhRBc42qIru_OyWkDbcDFv85Jsb4rVvfJ9nnzsFpuE9BCJcCyaarKWuTW7io0uQIjpdFZrnhTY620BAsSaqU4h8pIzjMl83xatW3knFwezx6Sl4N37-B_yr8GykMDE3F1JAbff4wYYrnvR99NP5WC5UIoa6ySv2aMXbY</recordid><startdate>20180503</startdate><enddate>20180503</enddate><creator>Zhuang, Ni</creator><creator>Yan, Yan</creator><creator>Chen, Si</creator><creator>Wang, Hanzi</creator><creator>Shen, Chunhua</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20180503</creationdate><title>Multi-label Learning Based Deep Transfer Neural Network for Facial Attribute Classification</title><author>Zhuang, Ni ; Yan, Yan ; Chen, Si ; Wang, Hanzi ; Shen, Chunhua</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a524-a06060bb99e199dbe874a833856741fcec453a9a3ac4411ab831164377ab859f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Artificial neural networks</topic><topic>Classification</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer vision</topic><topic>Face recognition</topic><topic>Learning</topic><topic>Neural networks</topic><topic>Weight</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhuang, Ni</creatorcontrib><creatorcontrib>Yan, Yan</creatorcontrib><creatorcontrib>Chen, Si</creatorcontrib><creatorcontrib>Wang, Hanzi</creatorcontrib><creatorcontrib>Shen, Chunhua</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection (ProQuest)</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhuang, Ni</au><au>Yan, Yan</au><au>Chen, Si</au><au>Wang, Hanzi</au><au>Shen, Chunhua</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multi-label Learning Based Deep Transfer Neural Network for Facial Attribute Classification</atitle><jtitle>arXiv.org</jtitle><date>2018-05-03</date><risdate>2018</risdate><eissn>2331-8422</eissn><abstract>Deep Neural Network (DNN) has recently achieved outstanding performance in a variety of computer vision tasks, including facial attribute classification. The great success of classifying facial attributes with DNN often relies on a massive amount of labelled data. However, in real-world applications, labelled data are only provided for some commonly used attributes (such as age, gender); whereas, unlabelled data are available for other attributes (such as attraction, hairline). To address the above problem, we propose a novel deep transfer neural network method based on multi-label learning for facial attribute classification, termed FMTNet, which consists of three sub-networks: the Face detection Network (FNet), the Multi-label learning Network (MNet) and the Transfer learning Network (TNet). Firstly, based on the Faster Region-based Convolutional Neural Network (Faster R-CNN), FNet is fine-tuned for face detection. Then, MNet is fine-tuned by FNet to predict multiple attributes with labelled data, where an effective loss weight scheme is developed to explicitly exploit the correlation between facial attributes based on attribute grouping. Finally, based on MNet, TNet is trained by taking advantage of unsupervised domain adaptation for unlabelled facial attribute classification. The three sub-networks are tightly coupled to perform effective facial attribute classification. A distinguishing characteristic of the proposed FMTNet method is that the three sub-networks (FNet, MNet and TNet) are constructed in a similar network structure. Extensive experimental results on challenging face datasets demonstrate the effectiveness of our proposed method compared with several state-of-the-art methods.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.1805.01282</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2018-05 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_1805_01282 |
source | arXiv.org; Free E- Journals |
subjects | Artificial neural networks Classification Computer Science - Computer Vision and Pattern Recognition Computer vision Face recognition Learning Neural networks Weight |
title | Multi-label Learning Based Deep Transfer Neural Network for Facial Attribute Classification |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T08%3A04%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multi-label%20Learning%20Based%20Deep%20Transfer%20Neural%20Network%20for%20Facial%20Attribute%20Classification&rft.jtitle=arXiv.org&rft.au=Zhuang,%20Ni&rft.date=2018-05-03&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.1805.01282&rft_dat=%3Cproquest_arxiv%3E2072249894%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2072249894&rft_id=info:pmid/&rfr_iscdi=true |