DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning
Neural predictors have shown great potential in the evaluation process of neural architecture search (NAS). However, current predictor-based approaches overlook the fact that training a predictor necessitates a considerable number of trained neural networks as the labeled training set, which is cost...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Zheng, Shenghe Wang, Hongzhi Mu, Tianyu |
description | Neural predictors have shown great potential in the evaluation process of
neural architecture search (NAS). However, current predictor-based approaches
overlook the fact that training a predictor necessitates a considerable number
of trained neural networks as the labeled training set, which is costly to
obtain. Therefore, the critical issue in utilizing predictors for NAS is to
train a high-performance predictor using as few trained neural networks as
possible. Although some methods attempt to address this problem through
unsupervised learning, they often result in inaccurate predictions. We argue
that the unsupervised tasks intended for the common graph data are too
challenging for neural networks, causing unsupervised training to be
susceptible to performance crashes in NAS. To address this issue, we propose a
Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP).
Our method simplifies the contrastive task by designing a novel curriculum to
enhance the stability of unlabeled training data distribution during
contrastive training. Specifically, we propose a scheduler that ranks the
training data according to the contrastive difficulty of each data and then
inputs them to the contrastive learner in order. This approach concentrates the
training data distribution and makes contrastive training more efficient. By
using our method, the contrastive learner incrementally learns feature
representations via unsupervised data on a smooth learning curve, avoiding
performance crashes that may occur with excessively variable training data
distributions. We experimentally demonstrate that DCLP has high accuracy and
efficiency compared with existing predictors, and shows promising potential to
discover superior architectures in various search spaces when combined with
search strategies. Our code is available at:
https://github.com/Zhengsh123/DCLP. |
doi_str_mv | 10.48550/arxiv.2302.13020 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2302_13020</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2302_13020</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-f96d875e102ca66f43b1c3a1a43d5711b1f46f77273c079cd7dbe99aff7944fa3</originalsourceid><addsrcrecordid>eNotz71OwzAYhWEvDKjlApjwDSTYsWPX3SrzU6QIOnSPvviHWkpT9NUucPdA6XLe7UgPIbec1XLRtuwe8Cud6kawpua_w67J-sF2myV9DQVhpCt0u5SDywUD3WDwyeUD0s-Ud9QWxOTKWPbUHqaMcMzpFGgXAKc0vc_JVYTxGG4unZHt0-PWrqvu7fnFrroKlGZVNMovdBs4axwoFaUYuBPAQQrfas4HHqWKWjdaOKaN89oPwRiIURspI4gZufu_PVP6D0x7wO_-j9SfSeIHeH1Gww</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</title><source>arXiv.org</source><creator>Zheng, Shenghe ; Wang, Hongzhi ; Mu, Tianyu</creator><creatorcontrib>Zheng, Shenghe ; Wang, Hongzhi ; Mu, Tianyu</creatorcontrib><description>Neural predictors have shown great potential in the evaluation process of
neural architecture search (NAS). However, current predictor-based approaches
overlook the fact that training a predictor necessitates a considerable number
of trained neural networks as the labeled training set, which is costly to
obtain. Therefore, the critical issue in utilizing predictors for NAS is to
train a high-performance predictor using as few trained neural networks as
possible. Although some methods attempt to address this problem through
unsupervised learning, they often result in inaccurate predictions. We argue
that the unsupervised tasks intended for the common graph data are too
challenging for neural networks, causing unsupervised training to be
susceptible to performance crashes in NAS. To address this issue, we propose a
Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP).
Our method simplifies the contrastive task by designing a novel curriculum to
enhance the stability of unlabeled training data distribution during
contrastive training. Specifically, we propose a scheduler that ranks the
training data according to the contrastive difficulty of each data and then
inputs them to the contrastive learner in order. This approach concentrates the
training data distribution and makes contrastive training more efficient. By
using our method, the contrastive learner incrementally learns feature
representations via unsupervised data on a smooth learning curve, avoiding
performance crashes that may occur with excessively variable training data
distributions. We experimentally demonstrate that DCLP has high accuracy and
efficiency compared with existing predictors, and shows promising potential to
discover superior architectures in various search spaces when combined with
search strategies. Our code is available at:
https://github.com/Zhengsh123/DCLP.</description><identifier>DOI: 10.48550/arxiv.2302.13020</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2023-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2302.13020$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2302.13020$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zheng, Shenghe</creatorcontrib><creatorcontrib>Wang, Hongzhi</creatorcontrib><creatorcontrib>Mu, Tianyu</creatorcontrib><title>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</title><description>Neural predictors have shown great potential in the evaluation process of
neural architecture search (NAS). However, current predictor-based approaches
overlook the fact that training a predictor necessitates a considerable number
of trained neural networks as the labeled training set, which is costly to
obtain. Therefore, the critical issue in utilizing predictors for NAS is to
train a high-performance predictor using as few trained neural networks as
possible. Although some methods attempt to address this problem through
unsupervised learning, they often result in inaccurate predictions. We argue
that the unsupervised tasks intended for the common graph data are too
challenging for neural networks, causing unsupervised training to be
susceptible to performance crashes in NAS. To address this issue, we propose a
Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP).
Our method simplifies the contrastive task by designing a novel curriculum to
enhance the stability of unlabeled training data distribution during
contrastive training. Specifically, we propose a scheduler that ranks the
training data according to the contrastive difficulty of each data and then
inputs them to the contrastive learner in order. This approach concentrates the
training data distribution and makes contrastive training more efficient. By
using our method, the contrastive learner incrementally learns feature
representations via unsupervised data on a smooth learning curve, avoiding
performance crashes that may occur with excessively variable training data
distributions. We experimentally demonstrate that DCLP has high accuracy and
efficiency compared with existing predictors, and shows promising potential to
discover superior architectures in various search spaces when combined with
search strategies. Our code is available at:
https://github.com/Zhengsh123/DCLP.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz71OwzAYhWEvDKjlApjwDSTYsWPX3SrzU6QIOnSPvviHWkpT9NUucPdA6XLe7UgPIbec1XLRtuwe8Cud6kawpua_w67J-sF2myV9DQVhpCt0u5SDywUD3WDwyeUD0s-Ud9QWxOTKWPbUHqaMcMzpFGgXAKc0vc_JVYTxGG4unZHt0-PWrqvu7fnFrroKlGZVNMovdBs4axwoFaUYuBPAQQrfas4HHqWKWjdaOKaN89oPwRiIURspI4gZufu_PVP6D0x7wO_-j9SfSeIHeH1Gww</recordid><startdate>20230225</startdate><enddate>20230225</enddate><creator>Zheng, Shenghe</creator><creator>Wang, Hongzhi</creator><creator>Mu, Tianyu</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230225</creationdate><title>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</title><author>Zheng, Shenghe ; Wang, Hongzhi ; Mu, Tianyu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-f96d875e102ca66f43b1c3a1a43d5711b1f46f77273c079cd7dbe99aff7944fa3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Zheng, Shenghe</creatorcontrib><creatorcontrib>Wang, Hongzhi</creatorcontrib><creatorcontrib>Mu, Tianyu</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zheng, Shenghe</au><au>Wang, Hongzhi</au><au>Mu, Tianyu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning</atitle><date>2023-02-25</date><risdate>2023</risdate><abstract>Neural predictors have shown great potential in the evaluation process of
neural architecture search (NAS). However, current predictor-based approaches
overlook the fact that training a predictor necessitates a considerable number
of trained neural networks as the labeled training set, which is costly to
obtain. Therefore, the critical issue in utilizing predictors for NAS is to
train a high-performance predictor using as few trained neural networks as
possible. Although some methods attempt to address this problem through
unsupervised learning, they often result in inaccurate predictions. We argue
that the unsupervised tasks intended for the common graph data are too
challenging for neural networks, causing unsupervised training to be
susceptible to performance crashes in NAS. To address this issue, we propose a
Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP).
Our method simplifies the contrastive task by designing a novel curriculum to
enhance the stability of unlabeled training data distribution during
contrastive training. Specifically, we propose a scheduler that ranks the
training data according to the contrastive difficulty of each data and then
inputs them to the contrastive learner in order. This approach concentrates the
training data distribution and makes contrastive training more efficient. By
using our method, the contrastive learner incrementally learns feature
representations via unsupervised data on a smooth learning curve, avoiding
performance crashes that may occur with excessively variable training data
distributions. We experimentally demonstrate that DCLP has high accuracy and
efficiency compared with existing predictors, and shows promising potential to
discover superior architectures in various search spaces when combined with
search strategies. Our code is available at:
https://github.com/Zhengsh123/DCLP.</abstract><doi>10.48550/arxiv.2302.13020</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2302.13020 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2302_13020 |
source | arXiv.org |
subjects | Computer Science - Learning |
title | DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T20%3A58%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DCLP:%20Neural%20Architecture%20Predictor%20with%20Curriculum%20Contrastive%20Learning&rft.au=Zheng,%20Shenghe&rft.date=2023-02-25&rft_id=info:doi/10.48550/arxiv.2302.13020&rft_dat=%3Carxiv_GOX%3E2302_13020%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |