Multitask-Guided Deep Clustering With Boundary Adaptation

Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initia...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transaction on neural networks and learning systems 2024-05, Vol.35 (5), p.6089-6102
Hauptverfasser: Zhang, Xiaobo, Wang, Tao, Zhao, Xiaole, Wen, Dengmin, Zhai, Donghai
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 6102
container_issue 5
container_start_page 6089
container_title IEEE transaction on neural networks and learning systems
container_volume 35
creator Zhang, Xiaobo
Wang, Tao
Zhao, Xiaole
Wen, Dengmin
Zhai, Donghai
description Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.
doi_str_mv 10.1109/TNNLS.2023.3307126
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_2860405713</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10236448</ieee_id><sourcerecordid>2860405713</sourcerecordid><originalsourceid>FETCH-LOGICAL-c303t-8c2068c357728fe7d97406f301792a98c5901facf74d8f0a4ef113eb2eeaca503</originalsourceid><addsrcrecordid>eNpdkMFOwzAMhiMEYmjsBRBClbhw6XCSNkmPY8BAGuPAENyirHWgo2tL0x54ezI2JoQv9uHzL_sj5ITCkFJILuez2fRpyIDxIecgKRN75IhRwULGldrfzfK1RwbOLcGXgFhEySHpcSliGil5RJKHrmjz1riPcNLlGWbBNWIdjIvOtdjk5VvwkrfvwVXVlZlpvoJRZurWtHlVHpMDawqHg23vk-fbm_n4Lpw-Tu7Ho2mYcuBtqFIGQqU8lpIpizJLZATCcqAyYSZRaZwAtSa1MsqUBROhpZTjgiGa1MTA--Rik1s31WeHrtWr3KVYFKbEqnOaKQERxJJyj57_Q5dV15T-Os3BRwGXQD3FNlTaVM41aHXd5Cv_nKag1271j1u9dqu3bv3S2Ta6W6ww2638mvTA6QbIEfFPIuMiihT_Bp5Te6U</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3050303701</pqid></control><display><type>article</type><title>Multitask-Guided Deep Clustering With Boundary Adaptation</title><source>IEEE Electronic Library (IEL)</source><creator>Zhang, Xiaobo ; Wang, Tao ; Zhao, Xiaole ; Wen, Dengmin ; Zhai, Donghai</creator><creatorcontrib>Zhang, Xiaobo ; Wang, Tao ; Zhao, Xiaole ; Wen, Dengmin ; Zhai, Donghai</creatorcontrib><description>Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.</description><identifier>ISSN: 2162-237X</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2023.3307126</identifier><identifier>PMID: 37651487</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Adaptation ; Algorithms ; Artificial neural networks ; Boundary adaptation ; Clustering ; Clustering algorithms ; Convolutional neural networks ; Correlation ; Data augmentation ; Data mining ; Data models ; Datasets ; deep learning ; explainable method ; Feature extraction ; Learning ; Machine learning ; Mathematical models ; multitask learning ; Neural networks ; Parameter sensitivity ; Parameters ; Principles ; Reconfiguration ; Task analysis ; Visualization</subject><ispartof>IEEE transaction on neural networks and learning systems, 2024-05, Vol.35 (5), p.6089-6102</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c303t-8c2068c357728fe7d97406f301792a98c5901facf74d8f0a4ef113eb2eeaca503</cites><orcidid>0000-0002-6598-0519 ; 0000-0003-0100-2414 ; 0000-0001-8396-5710</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10236448$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>315,782,786,798,27931,27932,54765</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10236448$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37651487$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhang, Xiaobo</creatorcontrib><creatorcontrib>Wang, Tao</creatorcontrib><creatorcontrib>Zhao, Xiaole</creatorcontrib><creatorcontrib>Wen, Dengmin</creatorcontrib><creatorcontrib>Zhai, Donghai</creatorcontrib><title>Multitask-Guided Deep Clustering With Boundary Adaptation</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.</description><subject>Adaptation</subject><subject>Algorithms</subject><subject>Artificial neural networks</subject><subject>Boundary adaptation</subject><subject>Clustering</subject><subject>Clustering algorithms</subject><subject>Convolutional neural networks</subject><subject>Correlation</subject><subject>Data augmentation</subject><subject>Data mining</subject><subject>Data models</subject><subject>Datasets</subject><subject>deep learning</subject><subject>explainable method</subject><subject>Feature extraction</subject><subject>Learning</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>multitask learning</subject><subject>Neural networks</subject><subject>Parameter sensitivity</subject><subject>Parameters</subject><subject>Principles</subject><subject>Reconfiguration</subject><subject>Task analysis</subject><subject>Visualization</subject><issn>2162-237X</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpdkMFOwzAMhiMEYmjsBRBClbhw6XCSNkmPY8BAGuPAENyirHWgo2tL0x54ezI2JoQv9uHzL_sj5ITCkFJILuez2fRpyIDxIecgKRN75IhRwULGldrfzfK1RwbOLcGXgFhEySHpcSliGil5RJKHrmjz1riPcNLlGWbBNWIdjIvOtdjk5VvwkrfvwVXVlZlpvoJRZurWtHlVHpMDawqHg23vk-fbm_n4Lpw-Tu7Ho2mYcuBtqFIGQqU8lpIpizJLZATCcqAyYSZRaZwAtSa1MsqUBROhpZTjgiGa1MTA--Rik1s31WeHrtWr3KVYFKbEqnOaKQERxJJyj57_Q5dV15T-Os3BRwGXQD3FNlTaVM41aHXd5Cv_nKag1271j1u9dqu3bv3S2Ta6W6ww2638mvTA6QbIEfFPIuMiihT_Bp5Te6U</recordid><startdate>20240501</startdate><enddate>20240501</enddate><creator>Zhang, Xiaobo</creator><creator>Wang, Tao</creator><creator>Zhao, Xiaole</creator><creator>Wen, Dengmin</creator><creator>Zhai, Donghai</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QP</scope><scope>7QQ</scope><scope>7QR</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7TK</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-6598-0519</orcidid><orcidid>https://orcid.org/0000-0003-0100-2414</orcidid><orcidid>https://orcid.org/0000-0001-8396-5710</orcidid></search><sort><creationdate>20240501</creationdate><title>Multitask-Guided Deep Clustering With Boundary Adaptation</title><author>Zhang, Xiaobo ; Wang, Tao ; Zhao, Xiaole ; Wen, Dengmin ; Zhai, Donghai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c303t-8c2068c357728fe7d97406f301792a98c5901facf74d8f0a4ef113eb2eeaca503</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Adaptation</topic><topic>Algorithms</topic><topic>Artificial neural networks</topic><topic>Boundary adaptation</topic><topic>Clustering</topic><topic>Clustering algorithms</topic><topic>Convolutional neural networks</topic><topic>Correlation</topic><topic>Data augmentation</topic><topic>Data mining</topic><topic>Data models</topic><topic>Datasets</topic><topic>deep learning</topic><topic>explainable method</topic><topic>Feature extraction</topic><topic>Learning</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>multitask learning</topic><topic>Neural networks</topic><topic>Parameter sensitivity</topic><topic>Parameters</topic><topic>Principles</topic><topic>Reconfiguration</topic><topic>Task analysis</topic><topic>Visualization</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Xiaobo</creatorcontrib><creatorcontrib>Wang, Tao</creatorcontrib><creatorcontrib>Zhao, Xiaole</creatorcontrib><creatorcontrib>Wen, Dengmin</creatorcontrib><creatorcontrib>Zhai, Donghai</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Calcium &amp; Calcified Tissue Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhang, Xiaobo</au><au>Wang, Tao</au><au>Zhao, Xiaole</au><au>Wen, Dengmin</au><au>Zhai, Donghai</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multitask-Guided Deep Clustering With Boundary Adaptation</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2024-05-01</date><risdate>2024</risdate><volume>35</volume><issue>5</issue><spage>6089</spage><epage>6102</epage><pages>6089-6102</pages><issn>2162-237X</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>Multitask learning uses external knowledge to improve internal clustering and single-task learning. Existing multitask learning algorithms mostly use shallow-level correlation to aid judgment, and the boundary factors on high-dimensional datasets often lead algorithms to poor performance. The initial parameters of these algorithms cause the border samples to fall into a local optimal solution. In this study, a multitask-guided deep clustering (DC) with boundary adaptation (MTDC-BA) based on a convolutional neural network autoencoder (CNN-AE) is proposed. In the first stage, dubbed multitask pretraining (M-train), we construct an autoencoder (AE) named CNN-AE using the DenseNet-like structure, which performs deep feature extraction and stores captured multitask knowledge into model parameters. In the second phase, the parameters of the M-train are shared for CNN-AE, and clustering results are obtained by deep features, which is termed as single-task fitting (S-fit). To eliminate the boundary effect, we use data augmentation and improved self-paced learning to construct the boundary adaptation. We integrate boundary adaptors into the M-train and S-fit stages appropriately. The interpretability of MTDC-BA is accomplished by data transformation. The model relies on the principle that features become important as the reconfiguration loss decreases. Experiments on a series of typical datasets confirm the performance of the proposed MTDC-BA. Compared with other traditional clustering methods, including single-task DC algorithms and the latest multitask clustering algorithms, our MTDC-BA achieves better clustering performance with higher computational efficiency. Deep features clustering results demonstrate the stability of MTDC-BA by visualization and convergence verification. Through the visualization experiment, we explain and analyze the whole model data input and the middle characteristic layer. Further understanding of the principle of MTDC-BA. Through additional experiments, we know that the proposed MTDC-BA is efficient in the use of multitask knowledge. Finally, we carry out sensitivity experiments on the hyper-parameters to verify their optimal performance.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>37651487</pmid><doi>10.1109/TNNLS.2023.3307126</doi><tpages>14</tpages><orcidid>https://orcid.org/0000-0002-6598-0519</orcidid><orcidid>https://orcid.org/0000-0003-0100-2414</orcidid><orcidid>https://orcid.org/0000-0001-8396-5710</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 2162-237X
ispartof IEEE transaction on neural networks and learning systems, 2024-05, Vol.35 (5), p.6089-6102
issn 2162-237X
2162-2388
language eng
recordid cdi_proquest_miscellaneous_2860405713
source IEEE Electronic Library (IEL)
subjects Adaptation
Algorithms
Artificial neural networks
Boundary adaptation
Clustering
Clustering algorithms
Convolutional neural networks
Correlation
Data augmentation
Data mining
Data models
Datasets
deep learning
explainable method
Feature extraction
Learning
Machine learning
Mathematical models
multitask learning
Neural networks
Parameter sensitivity
Parameters
Principles
Reconfiguration
Task analysis
Visualization
title Multitask-Guided Deep Clustering With Boundary Adaptation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-04T02%3A46%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multitask-Guided%20Deep%20Clustering%20With%20Boundary%20Adaptation&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Zhang,%20Xiaobo&rft.date=2024-05-01&rft.volume=35&rft.issue=5&rft.spage=6089&rft.epage=6102&rft.pages=6089-6102&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2023.3307126&rft_dat=%3Cproquest_RIE%3E2860405713%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3050303701&rft_id=info:pmid/37651487&rft_ieee_id=10236448&rfr_iscdi=true