NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing

Neural Architecture Search (NAS) is a promising and rapidly evolving research area. Training a large number of neural networks requires an exceptional amount of computational power, which makes NAS unreachable for those researchers who have limited or no access to high-performance clusters and super...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Klyuchnikov, Nikita, Trofimov, Ilya, Artemova, Ekaterina, Salnikov, Mikhail, Fedorov, Maxim, Burnaev, Evgeny
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Klyuchnikov, Nikita
Trofimov, Ilya
Artemova, Ekaterina
Salnikov, Mikhail
Fedorov, Maxim
Burnaev, Evgeny
description Neural Architecture Search (NAS) is a promising and rapidly evolving research area. Training a large number of neural networks requires an exceptional amount of computational power, which makes NAS unreachable for those researchers who have limited or no access to high-performance clusters and supercomputers. A few benchmarks with precomputed neural architectures performances have been recently introduced to overcome this problem and ensure more reproducible experiments. However, these benchmarks are only for the computer vision domain and, thus, are built from the image datasets and convolution-derived architectures. In this work, we step outside the computer vision domain by leveraging the language modeling task, which is the core of natural language processing (NLP). Our main contribution is as follows: we have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it; we have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation; finally, we have tested several NAS algorithms to demonstrate how the precomputed results can be utilized. We believe that our results have high potential of usage for both NAS and NLP communities.
doi_str_mv 10.48550/arxiv.2006.07116
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2006_07116</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2006_07116</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-6878122a5afe569ee29c7c15ecca68a02f7e90cbe7220fb1079a6ea4560536663</originalsourceid><addsrcrecordid>eNotj81OwzAQhH3hgFoegBN-AYe1U6-T3kLFnxSFSuk92lrrNKKkyEkQvD0l5TQa6dNoPiFuNSSrzFq4p_jdfSUGABNwWuO1qKuiVg_c-4Oqyu1aVjxFOsoi-kM3sh-nyLJmOlc5Ux8U32U4RVnROJMl9e1ELcttPHkehq5vl-Iq0HHgm_9ciN3T427zosq359dNUSpChwozl2ljyFJgizmzyb3z2rL3hBmBCY5z8Ht2xkDYa3A5IdPKItgUEdOFuLvMzlbNZ-zO536aP7tmtkt_AWLVSW4</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing</title><source>arXiv.org</source><creator>Klyuchnikov, Nikita ; Trofimov, Ilya ; Artemova, Ekaterina ; Salnikov, Mikhail ; Fedorov, Maxim ; Burnaev, Evgeny</creator><creatorcontrib>Klyuchnikov, Nikita ; Trofimov, Ilya ; Artemova, Ekaterina ; Salnikov, Mikhail ; Fedorov, Maxim ; Burnaev, Evgeny</creatorcontrib><description>Neural Architecture Search (NAS) is a promising and rapidly evolving research area. Training a large number of neural networks requires an exceptional amount of computational power, which makes NAS unreachable for those researchers who have limited or no access to high-performance clusters and supercomputers. A few benchmarks with precomputed neural architectures performances have been recently introduced to overcome this problem and ensure more reproducible experiments. However, these benchmarks are only for the computer vision domain and, thus, are built from the image datasets and convolution-derived architectures. In this work, we step outside the computer vision domain by leveraging the language modeling task, which is the core of natural language processing (NLP). Our main contribution is as follows: we have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it; we have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation; finally, we have tested several NAS algorithms to demonstrate how the precomputed results can be utilized. We believe that our results have high potential of usage for both NAS and NLP communities.</description><identifier>DOI: 10.48550/arxiv.2006.07116</identifier><language>eng</language><subject>Computer Science - Computation and Language ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2020-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2006.07116$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2006.07116$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Klyuchnikov, Nikita</creatorcontrib><creatorcontrib>Trofimov, Ilya</creatorcontrib><creatorcontrib>Artemova, Ekaterina</creatorcontrib><creatorcontrib>Salnikov, Mikhail</creatorcontrib><creatorcontrib>Fedorov, Maxim</creatorcontrib><creatorcontrib>Burnaev, Evgeny</creatorcontrib><title>NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing</title><description>Neural Architecture Search (NAS) is a promising and rapidly evolving research area. Training a large number of neural networks requires an exceptional amount of computational power, which makes NAS unreachable for those researchers who have limited or no access to high-performance clusters and supercomputers. A few benchmarks with precomputed neural architectures performances have been recently introduced to overcome this problem and ensure more reproducible experiments. However, these benchmarks are only for the computer vision domain and, thus, are built from the image datasets and convolution-derived architectures. In this work, we step outside the computer vision domain by leveraging the language modeling task, which is the core of natural language processing (NLP). Our main contribution is as follows: we have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it; we have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation; finally, we have tested several NAS algorithms to demonstrate how the precomputed results can be utilized. We believe that our results have high potential of usage for both NAS and NLP communities.</description><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj81OwzAQhH3hgFoegBN-AYe1U6-T3kLFnxSFSuk92lrrNKKkyEkQvD0l5TQa6dNoPiFuNSSrzFq4p_jdfSUGABNwWuO1qKuiVg_c-4Oqyu1aVjxFOsoi-kM3sh-nyLJmOlc5Ux8U32U4RVnROJMl9e1ELcttPHkehq5vl-Iq0HHgm_9ciN3T427zosq359dNUSpChwozl2ljyFJgizmzyb3z2rL3hBmBCY5z8Ht2xkDYa3A5IdPKItgUEdOFuLvMzlbNZ-zO536aP7tmtkt_AWLVSW4</recordid><startdate>20200612</startdate><enddate>20200612</enddate><creator>Klyuchnikov, Nikita</creator><creator>Trofimov, Ilya</creator><creator>Artemova, Ekaterina</creator><creator>Salnikov, Mikhail</creator><creator>Fedorov, Maxim</creator><creator>Burnaev, Evgeny</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20200612</creationdate><title>NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing</title><author>Klyuchnikov, Nikita ; Trofimov, Ilya ; Artemova, Ekaterina ; Salnikov, Mikhail ; Fedorov, Maxim ; Burnaev, Evgeny</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-6878122a5afe569ee29c7c15ecca68a02f7e90cbe7220fb1079a6ea4560536663</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Klyuchnikov, Nikita</creatorcontrib><creatorcontrib>Trofimov, Ilya</creatorcontrib><creatorcontrib>Artemova, Ekaterina</creatorcontrib><creatorcontrib>Salnikov, Mikhail</creatorcontrib><creatorcontrib>Fedorov, Maxim</creatorcontrib><creatorcontrib>Burnaev, Evgeny</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Klyuchnikov, Nikita</au><au>Trofimov, Ilya</au><au>Artemova, Ekaterina</au><au>Salnikov, Mikhail</au><au>Fedorov, Maxim</au><au>Burnaev, Evgeny</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing</atitle><date>2020-06-12</date><risdate>2020</risdate><abstract>Neural Architecture Search (NAS) is a promising and rapidly evolving research area. Training a large number of neural networks requires an exceptional amount of computational power, which makes NAS unreachable for those researchers who have limited or no access to high-performance clusters and supercomputers. A few benchmarks with precomputed neural architectures performances have been recently introduced to overcome this problem and ensure more reproducible experiments. However, these benchmarks are only for the computer vision domain and, thus, are built from the image datasets and convolution-derived architectures. In this work, we step outside the computer vision domain by leveraging the language modeling task, which is the core of natural language processing (NLP). Our main contribution is as follows: we have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it; we have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation; finally, we have tested several NAS algorithms to demonstrate how the precomputed results can be utilized. We believe that our results have high potential of usage for both NAS and NLP communities.</abstract><doi>10.48550/arxiv.2006.07116</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2006.07116
ispartof
issn
language eng
recordid cdi_arxiv_primary_2006_07116
source arXiv.org
subjects Computer Science - Computation and Language
Computer Science - Learning
Statistics - Machine Learning
title NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T13%3A54%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=NAS-Bench-NLP:%20Neural%20Architecture%20Search%20Benchmark%20for%20Natural%20Language%20Processing&rft.au=Klyuchnikov,%20Nikita&rft.date=2020-06-12&rft_id=info:doi/10.48550/arxiv.2006.07116&rft_dat=%3Carxiv_GOX%3E2006_07116%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true