DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning

Neural predictors have shown great potential in the evaluation process of neural architecture search (NAS). However, current predictor-based approaches overlook the fact that training a predictor necessitates a considerable number of trained neural networks as the labeled training set, which is cost...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Zheng, Shenghe, Wang, Hongzhi, Mu, Tianyu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Zheng, Shenghe
Wang, Hongzhi
Mu, Tianyu
description Neural predictors have shown great potential in the evaluation process of neural architecture search (NAS). However, current predictor-based approaches overlook the fact that training a predictor necessitates a considerable number of trained neural networks as the labeled training set, which is costly to obtain. Therefore, the critical issue in utilizing predictors for NAS is to train a high-performance predictor using as few trained neural networks as possible. Although some methods attempt to address this problem through unsupervised learning, they often result in inaccurate predictions. We argue that the unsupervised tasks intended for the common graph data are too challenging for neural networks, causing unsupervised training to be susceptible to performance crashes in NAS. To address this issue, we propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP). Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution during contrastive training. Specifically, we propose a scheduler that ranks the training data according to the contrastive difficulty of each data and then inputs them to the contrastive learner in order. This approach concentrates the training data distribution and makes contrastive training more efficient. By using our method, the contrastive learner incrementally learns feature representations via unsupervised data on a smooth learning curve, avoiding performance crashes that may occur with excessively variable training data distributions. We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors, and shows promising potential to discover superior architectures in various search spaces when combined with search strategies. Our code is available at: https://github.com/Zhengsh123/DCLP.
doi_str_mv 10.48550/arxiv.2302.13020
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2302_13020</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2302_13020</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-f96d875e102ca66f43b1c3a1a43d5711b1f46f77273c079cd7dbe99aff7944fa3</originalsourceid><addsrcrecordid>eNotz71OwzAYhWEvDKjlApjwDSTYsWPX3SrzU6QIOnSPvviHWkpT9NUucPdA6XLe7UgPIbec1XLRtuwe8Cud6kawpua_w67J-sF2myV9DQVhpCt0u5SDywUD3WDwyeUD0s-Ud9QWxOTKWPbUHqaMcMzpFGgXAKc0vc_JVYTxGG4unZHt0-PWrqvu7fnFrroKlGZVNMovdBs4axwoFaUYuBPAQQrfas4HHqWKWjdaOKaN89oPwRiIURspI4gZufu_PVP6D0x7wO_-j9SfSeIHeH1Gww</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</title><source>arXiv.org</source><creator>Zheng, Shenghe ; Wang, Hongzhi ; Mu, Tianyu</creator><creatorcontrib>Zheng, Shenghe ; Wang, Hongzhi ; Mu, Tianyu</creatorcontrib><description>Neural predictors have shown great potential in the evaluation process of neural architecture search (NAS). However, current predictor-based approaches overlook the fact that training a predictor necessitates a considerable number of trained neural networks as the labeled training set, which is costly to obtain. Therefore, the critical issue in utilizing predictors for NAS is to train a high-performance predictor using as few trained neural networks as possible. Although some methods attempt to address this problem through unsupervised learning, they often result in inaccurate predictions. We argue that the unsupervised tasks intended for the common graph data are too challenging for neural networks, causing unsupervised training to be susceptible to performance crashes in NAS. To address this issue, we propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP). Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution during contrastive training. Specifically, we propose a scheduler that ranks the training data according to the contrastive difficulty of each data and then inputs them to the contrastive learner in order. This approach concentrates the training data distribution and makes contrastive training more efficient. By using our method, the contrastive learner incrementally learns feature representations via unsupervised data on a smooth learning curve, avoiding performance crashes that may occur with excessively variable training data distributions. We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors, and shows promising potential to discover superior architectures in various search spaces when combined with search strategies. Our code is available at: https://github.com/Zhengsh123/DCLP.</description><identifier>DOI: 10.48550/arxiv.2302.13020</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2023-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2302.13020$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2302.13020$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zheng, Shenghe</creatorcontrib><creatorcontrib>Wang, Hongzhi</creatorcontrib><creatorcontrib>Mu, Tianyu</creatorcontrib><title>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</title><description>Neural predictors have shown great potential in the evaluation process of neural architecture search (NAS). However, current predictor-based approaches overlook the fact that training a predictor necessitates a considerable number of trained neural networks as the labeled training set, which is costly to obtain. Therefore, the critical issue in utilizing predictors for NAS is to train a high-performance predictor using as few trained neural networks as possible. Although some methods attempt to address this problem through unsupervised learning, they often result in inaccurate predictions. We argue that the unsupervised tasks intended for the common graph data are too challenging for neural networks, causing unsupervised training to be susceptible to performance crashes in NAS. To address this issue, we propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP). Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution during contrastive training. Specifically, we propose a scheduler that ranks the training data according to the contrastive difficulty of each data and then inputs them to the contrastive learner in order. This approach concentrates the training data distribution and makes contrastive training more efficient. By using our method, the contrastive learner incrementally learns feature representations via unsupervised data on a smooth learning curve, avoiding performance crashes that may occur with excessively variable training data distributions. We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors, and shows promising potential to discover superior architectures in various search spaces when combined with search strategies. Our code is available at: https://github.com/Zhengsh123/DCLP.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz71OwzAYhWEvDKjlApjwDSTYsWPX3SrzU6QIOnSPvviHWkpT9NUucPdA6XLe7UgPIbec1XLRtuwe8Cud6kawpua_w67J-sF2myV9DQVhpCt0u5SDywUD3WDwyeUD0s-Ud9QWxOTKWPbUHqaMcMzpFGgXAKc0vc_JVYTxGG4unZHt0-PWrqvu7fnFrroKlGZVNMovdBs4axwoFaUYuBPAQQrfas4HHqWKWjdaOKaN89oPwRiIURspI4gZufu_PVP6D0x7wO_-j9SfSeIHeH1Gww</recordid><startdate>20230225</startdate><enddate>20230225</enddate><creator>Zheng, Shenghe</creator><creator>Wang, Hongzhi</creator><creator>Mu, Tianyu</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230225</creationdate><title>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</title><author>Zheng, Shenghe ; Wang, Hongzhi ; Mu, Tianyu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-f96d875e102ca66f43b1c3a1a43d5711b1f46f77273c079cd7dbe99aff7944fa3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Zheng, Shenghe</creatorcontrib><creatorcontrib>Wang, Hongzhi</creatorcontrib><creatorcontrib>Mu, Tianyu</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zheng, Shenghe</au><au>Wang, Hongzhi</au><au>Mu, Tianyu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</atitle><date>2023-02-25</date><risdate>2023</risdate><abstract>Neural predictors have shown great potential in the evaluation process of neural architecture search (NAS). However, current predictor-based approaches overlook the fact that training a predictor necessitates a considerable number of trained neural networks as the labeled training set, which is costly to obtain. Therefore, the critical issue in utilizing predictors for NAS is to train a high-performance predictor using as few trained neural networks as possible. Although some methods attempt to address this problem through unsupervised learning, they often result in inaccurate predictions. We argue that the unsupervised tasks intended for the common graph data are too challenging for neural networks, causing unsupervised training to be susceptible to performance crashes in NAS. To address this issue, we propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP). Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution during contrastive training. Specifically, we propose a scheduler that ranks the training data according to the contrastive difficulty of each data and then inputs them to the contrastive learner in order. This approach concentrates the training data distribution and makes contrastive training more efficient. By using our method, the contrastive learner incrementally learns feature representations via unsupervised data on a smooth learning curve, avoiding performance crashes that may occur with excessively variable training data distributions. We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors, and shows promising potential to discover superior architectures in various search spaces when combined with search strategies. Our code is available at: https://github.com/Zhengsh123/DCLP.</abstract><doi>10.48550/arxiv.2302.13020</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2302.13020
ispartof
issn
language eng
recordid cdi_arxiv_primary_2302_13020
source arXiv.org
subjects Computer Science - Learning
title DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T20%3A58%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DCLP:%20Neural%20Architecture%20Predictor%20with%20Curriculum%20Contrastive%20Learning&rft.au=Zheng,%20Shenghe&rft.date=2023-02-25&rft_id=info:doi/10.48550/arxiv.2302.13020&rft_dat=%3Carxiv_GOX%3E2302_13020%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true