Multitask-Guided Deep Clustering With Boundary Adaptation
Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initia...
Gespeichert in:
Veröffentlicht in: | IEEE transaction on neural networks and learning systems 2024-05, Vol.35 (5), p.6089-6102 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 6102 |
---|---|
container_issue | 5 |
container_start_page | 6089 |
container_title | IEEE transaction on neural networks and learning systems |
container_volume | 35 |
creator | Zhang, Xiaobo Wang, Tao Zhao, Xiaole Wen, Dengmin Zhai, Donghai |
description | Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance. |
doi_str_mv | 10.1109/TNNLS.2023.3307126 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_2860405713</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10236448</ieee_id><sourcerecordid>2860405713</sourcerecordid><originalsourceid>FETCH-LOGICAL-c303t-8c2068c357728fe7d97406f301792a98c5901facf74d8f0a4ef113eb2eeaca503</originalsourceid><addsrcrecordid>eNpdkMFOwzAMhiMEYmjsBRBClbhw6XCSNkmPY8BAGuPAENyirHWgo2tL0x54ezI2JoQv9uHzL_sj5ITCkFJILuez2fRpyIDxIecgKRN75IhRwULGldrfzfK1RwbOLcGXgFhEySHpcSliGil5RJKHrmjz1riPcNLlGWbBNWIdjIvOtdjk5VvwkrfvwVXVlZlpvoJRZurWtHlVHpMDawqHg23vk-fbm_n4Lpw-Tu7Ho2mYcuBtqFIGQqU8lpIpizJLZATCcqAyYSZRaZwAtSa1MsqUBROhpZTjgiGa1MTA--Rik1s31WeHrtWr3KVYFKbEqnOaKQERxJJyj57_Q5dV15T-Os3BRwGXQD3FNlTaVM41aHXd5Cv_nKag1271j1u9dqu3bv3S2Ta6W6ww2638mvTA6QbIEfFPIuMiihT_Bp5Te6U</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3050303701</pqid></control><display><type>article</type><title>Multitask-Guided Deep Clustering With Boundary Adaptation</title><source>IEEE Electronic Library (IEL)</source><creator>Zhang, Xiaobo ; Wang, Tao ; Zhao, Xiaole ; Wen, Dengmin ; Zhai, Donghai</creator><creatorcontrib>Zhang, Xiaobo ; Wang, Tao ; Zhao, Xiaole ; Wen, Dengmin ; Zhai, Donghai</creatorcontrib><description>Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.</description><identifier>ISSN: 2162-237X</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2023.3307126</identifier><identifier>PMID: 37651487</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Adaptation ; Algorithms ; Artificial neural networks ; Boundary adaptation ; Clustering ; Clustering algorithms ; Convolutional neural networks ; Correlation ; Data augmentation ; Data mining ; Data models ; Datasets ; deep learning ; explainable method ; Feature extraction ; Learning ; Machine learning ; Mathematical models ; multitask learning ; Neural networks ; Parameter sensitivity ; Parameters ; Principles ; Reconfiguration ; Task analysis ; Visualization</subject><ispartof>IEEE transaction on neural networks and learning systems, 2024-05, Vol.35 (5), p.6089-6102</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c303t-8c2068c357728fe7d97406f301792a98c5901facf74d8f0a4ef113eb2eeaca503</cites><orcidid>0000-0002-6598-0519 ; 0000-0003-0100-2414 ; 0000-0001-8396-5710</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10236448$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>315,782,786,798,27931,27932,54765</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10236448$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37651487$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhang, Xiaobo</creatorcontrib><creatorcontrib>Wang, Tao</creatorcontrib><creatorcontrib>Zhao, Xiaole</creatorcontrib><creatorcontrib>Wen, Dengmin</creatorcontrib><creatorcontrib>Zhai, Donghai</creatorcontrib><title>Multitask-Guided Deep Clustering With Boundary Adaptation</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.</description><subject>Adaptation</subject><subject>Algorithms</subject><subject>Artificial neural networks</subject><subject>Boundary adaptation</subject><subject>Clustering</subject><subject>Clustering algorithms</subject><subject>Convolutional neural networks</subject><subject>Correlation</subject><subject>Data augmentation</subject><subject>Data mining</subject><subject>Data models</subject><subject>Datasets</subject><subject>deep learning</subject><subject>explainable method</subject><subject>Feature extraction</subject><subject>Learning</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>multitask learning</subject><subject>Neural networks</subject><subject>Parameter sensitivity</subject><subject>Parameters</subject><subject>Principles</subject><subject>Reconfiguration</subject><subject>Task analysis</subject><subject>Visualization</subject><issn>2162-237X</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpdkMFOwzAMhiMEYmjsBRBClbhw6XCSNkmPY8BAGuPAENyirHWgo2tL0x54ezI2JoQv9uHzL_sj5ITCkFJILuez2fRpyIDxIecgKRN75IhRwULGldrfzfK1RwbOLcGXgFhEySHpcSliGil5RJKHrmjz1riPcNLlGWbBNWIdjIvOtdjk5VvwkrfvwVXVlZlpvoJRZurWtHlVHpMDawqHg23vk-fbm_n4Lpw-Tu7Ho2mYcuBtqFIGQqU8lpIpizJLZATCcqAyYSZRaZwAtSa1MsqUBROhpZTjgiGa1MTA--Rik1s31WeHrtWr3KVYFKbEqnOaKQERxJJyj57_Q5dV15T-Os3BRwGXQD3FNlTaVM41aHXd5Cv_nKag1271j1u9dqu3bv3S2Ta6W6ww2638mvTA6QbIEfFPIuMiihT_Bp5Te6U</recordid><startdate>20240501</startdate><enddate>20240501</enddate><creator>Zhang, Xiaobo</creator><creator>Wang, Tao</creator><creator>Zhao, Xiaole</creator><creator>Wen, Dengmin</creator><creator>Zhai, Donghai</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QP</scope><scope>7QQ</scope><scope>7QR</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7TK</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-6598-0519</orcidid><orcidid>https://orcid.org/0000-0003-0100-2414</orcidid><orcidid>https://orcid.org/0000-0001-8396-5710</orcidid></search><sort><creationdate>20240501</creationdate><title>Multitask-Guided Deep Clustering With Boundary Adaptation</title><author>Zhang, Xiaobo ; Wang, Tao ; Zhao, Xiaole ; Wen, Dengmin ; Zhai, Donghai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c303t-8c2068c357728fe7d97406f301792a98c5901facf74d8f0a4ef113eb2eeaca503</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Adaptation</topic><topic>Algorithms</topic><topic>Artificial neural networks</topic><topic>Boundary adaptation</topic><topic>Clustering</topic><topic>Clustering algorithms</topic><topic>Convolutional neural networks</topic><topic>Correlation</topic><topic>Data augmentation</topic><topic>Data mining</topic><topic>Data models</topic><topic>Datasets</topic><topic>deep learning</topic><topic>explainable method</topic><topic>Feature extraction</topic><topic>Learning</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>multitask learning</topic><topic>Neural networks</topic><topic>Parameter sensitivity</topic><topic>Parameters</topic><topic>Principles</topic><topic>Reconfiguration</topic><topic>Task analysis</topic><topic>Visualization</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Xiaobo</creatorcontrib><creatorcontrib>Wang, Tao</creatorcontrib><creatorcontrib>Zhao, Xiaole</creatorcontrib><creatorcontrib>Wen, Dengmin</creatorcontrib><creatorcontrib>Zhai, Donghai</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Calcium & Calcified Tissue Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology & Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhang, Xiaobo</au><au>Wang, Tao</au><au>Zhao, Xiaole</au><au>Wen, Dengmin</au><au>Zhai, Donghai</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multitask-Guided Deep Clustering With Boundary Adaptation</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2024-05-01</date><risdate>2024</risdate><volume>35</volume><issue>5</issue><spage>6089</spage><epage>6102</epage><pages>6089-6102</pages><issn>2162-237X</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>37651487</pmid><doi>10.1109/TNNLS.2023.3307126</doi><tpages>14</tpages><orcidid>https://orcid.org/0000-0002-6598-0519</orcidid><orcidid>https://orcid.org/0000-0003-0100-2414</orcidid><orcidid>https://orcid.org/0000-0001-8396-5710</orcidid></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 2162-237X |
ispartof | IEEE transaction on neural networks and learning systems, 2024-05, Vol.35 (5), p.6089-6102 |
issn | 2162-237X 2162-2388 |
language | eng |
recordid | cdi_proquest_miscellaneous_2860405713 |
source | IEEE Electronic Library (IEL) |
subjects | Adaptation Algorithms Artificial neural networks Boundary adaptation Clustering Clustering algorithms Convolutional neural networks Correlation Data augmentation Data mining Data models Datasets deep learning explainable method Feature extraction Learning Machine learning Mathematical models multitask learning Neural networks Parameter sensitivity Parameters Principles Reconfiguration Task analysis Visualization |
title | Multitask-Guided Deep Clustering With Boundary Adaptation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-04T02%3A46%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multitask-Guided%20Deep%20Clustering%20With%20Boundary%20Adaptation&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Zhang,%20Xiaobo&rft.date=2024-05-01&rft.volume=35&rft.issue=5&rft.spage=6089&rft.epage=6102&rft.pages=6089-6102&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2023.3307126&rft_dat=%3Cproquest_RIE%3E2860405713%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3050303701&rft_id=info:pmid/37651487&rft_ieee_id=10236448&rfr_iscdi=true |