A Two-Phase Recall-and-Select Framework for Fast Model Selection

As the ubiquity of deep learning in various machine learning applications has amplified, a proliferation of neural network models has been trained and shared on public model repositories. In the context of a targeted machine learning assignment, utilizing an apt source model as a starting point typi...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Cui, Jianwei, Shi, Wenhang, Tao, Honglin, Lu, Wei, Du, Xiaoyong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Cui, Jianwei
Shi, Wenhang
Tao, Honglin
Lu, Wei
Du, Xiaoyong
description As the ubiquity of deep learning in various machine learning applications has amplified, a proliferation of neural network models has been trained and shared on public model repositories. In the context of a targeted machine learning assignment, utilizing an apt source model as a starting point typically outperforms the strategy of training from scratch, particularly with limited training data. Despite the investigation and development of numerous model selection strategies in prior work, the process remains time-consuming, especially given the ever-increasing scale of model repositories. In this paper, we propose a two-phase (coarse-recall and fine-selection) model selection framework, aiming to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets. Specifically, the coarse-recall phase clusters models showcasing similar training performances on benchmark datasets in an offline manner. A light-weight proxy score is subsequently computed between this model cluster and the target dataset, which serves to recall a significantly smaller subset of potential candidate models in a swift manner. In the following fine-selection phase, the final model is chosen by fine-tuning the recalled models on the target dataset with successive halving. To accelerate the process, the final fine-tuning performance of each potential model is predicted by mining the model's convergence trend on the benchmark datasets, which aids in filtering lower performance models more earlier during fine-tuning. Through extensive experimentation on tasks covering natural language processing and computer vision, it has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods. Our code is available at https://github.com/plasware/two-phase-selection.
doi_str_mv 10.48550/arxiv.2404.00069
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2404_00069</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2404_00069</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-7c1ce280216d255b691a7023c6f885963a0c0fa2e036e90494a21d6a2303668a3</originalsourceid><addsrcrecordid>eNotz81KAzEUBeBsupDWB3BlXiDTm99JdpbiqFBRdPbDNXMHB9OmZIrVt1dbVwfOgQMfY1cSKuOthSWWr_GzUgZMBQAuXLCbFW-PWTy_40T8hSKmJHDXi1dKFA-8KbilYy4ffMiFNzgd-GPuKfHzPubdgs0GTBNd_uectc1tu74Xm6e7h_VqI9DVQdRRRlIelHS9svbNBYk1KB3d4L0NTiNEGFARaEcBTDCoZO9Q6d_CedRzdn2-PRG6fRm3WL67P0p3ougf78hBYA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Two-Phase Recall-and-Select Framework for Fast Model Selection</title><source>arXiv.org</source><creator>Cui, Jianwei ; Shi, Wenhang ; Tao, Honglin ; Lu, Wei ; Du, Xiaoyong</creator><creatorcontrib>Cui, Jianwei ; Shi, Wenhang ; Tao, Honglin ; Lu, Wei ; Du, Xiaoyong</creatorcontrib><description>As the ubiquity of deep learning in various machine learning applications has amplified, a proliferation of neural network models has been trained and shared on public model repositories. In the context of a targeted machine learning assignment, utilizing an apt source model as a starting point typically outperforms the strategy of training from scratch, particularly with limited training data. Despite the investigation and development of numerous model selection strategies in prior work, the process remains time-consuming, especially given the ever-increasing scale of model repositories. In this paper, we propose a two-phase (coarse-recall and fine-selection) model selection framework, aiming to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets. Specifically, the coarse-recall phase clusters models showcasing similar training performances on benchmark datasets in an offline manner. A light-weight proxy score is subsequently computed between this model cluster and the target dataset, which serves to recall a significantly smaller subset of potential candidate models in a swift manner. In the following fine-selection phase, the final model is chosen by fine-tuning the recalled models on the target dataset with successive halving. To accelerate the process, the final fine-tuning performance of each potential model is predicted by mining the model's convergence trend on the benchmark datasets, which aids in filtering lower performance models more earlier during fine-tuning. Through extensive experimentation on tasks covering natural language processing and computer vision, it has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods. Our code is available at https://github.com/plasware/two-phase-selection.</description><identifier>DOI: 10.48550/arxiv.2404.00069</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2024-03</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2404.00069$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2404.00069$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Cui, Jianwei</creatorcontrib><creatorcontrib>Shi, Wenhang</creatorcontrib><creatorcontrib>Tao, Honglin</creatorcontrib><creatorcontrib>Lu, Wei</creatorcontrib><creatorcontrib>Du, Xiaoyong</creatorcontrib><title>A Two-Phase Recall-and-Select Framework for Fast Model Selection</title><description>As the ubiquity of deep learning in various machine learning applications has amplified, a proliferation of neural network models has been trained and shared on public model repositories. In the context of a targeted machine learning assignment, utilizing an apt source model as a starting point typically outperforms the strategy of training from scratch, particularly with limited training data. Despite the investigation and development of numerous model selection strategies in prior work, the process remains time-consuming, especially given the ever-increasing scale of model repositories. In this paper, we propose a two-phase (coarse-recall and fine-selection) model selection framework, aiming to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets. Specifically, the coarse-recall phase clusters models showcasing similar training performances on benchmark datasets in an offline manner. A light-weight proxy score is subsequently computed between this model cluster and the target dataset, which serves to recall a significantly smaller subset of potential candidate models in a swift manner. In the following fine-selection phase, the final model is chosen by fine-tuning the recalled models on the target dataset with successive halving. To accelerate the process, the final fine-tuning performance of each potential model is predicted by mining the model's convergence trend on the benchmark datasets, which aids in filtering lower performance models more earlier during fine-tuning. Through extensive experimentation on tasks covering natural language processing and computer vision, it has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods. Our code is available at https://github.com/plasware/two-phase-selection.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz81KAzEUBeBsupDWB3BlXiDTm99JdpbiqFBRdPbDNXMHB9OmZIrVt1dbVwfOgQMfY1cSKuOthSWWr_GzUgZMBQAuXLCbFW-PWTy_40T8hSKmJHDXi1dKFA-8KbilYy4ffMiFNzgd-GPuKfHzPubdgs0GTBNd_uectc1tu74Xm6e7h_VqI9DVQdRRRlIelHS9svbNBYk1KB3d4L0NTiNEGFARaEcBTDCoZO9Q6d_CedRzdn2-PRG6fRm3WL67P0p3ougf78hBYA</recordid><startdate>20240328</startdate><enddate>20240328</enddate><creator>Cui, Jianwei</creator><creator>Shi, Wenhang</creator><creator>Tao, Honglin</creator><creator>Lu, Wei</creator><creator>Du, Xiaoyong</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240328</creationdate><title>A Two-Phase Recall-and-Select Framework for Fast Model Selection</title><author>Cui, Jianwei ; Shi, Wenhang ; Tao, Honglin ; Lu, Wei ; Du, Xiaoyong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-7c1ce280216d255b691a7023c6f885963a0c0fa2e036e90494a21d6a2303668a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Cui, Jianwei</creatorcontrib><creatorcontrib>Shi, Wenhang</creatorcontrib><creatorcontrib>Tao, Honglin</creatorcontrib><creatorcontrib>Lu, Wei</creatorcontrib><creatorcontrib>Du, Xiaoyong</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Cui, Jianwei</au><au>Shi, Wenhang</au><au>Tao, Honglin</au><au>Lu, Wei</au><au>Du, Xiaoyong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Two-Phase Recall-and-Select Framework for Fast Model Selection</atitle><date>2024-03-28</date><risdate>2024</risdate><abstract>As the ubiquity of deep learning in various machine learning applications has amplified, a proliferation of neural network models has been trained and shared on public model repositories. In the context of a targeted machine learning assignment, utilizing an apt source model as a starting point typically outperforms the strategy of training from scratch, particularly with limited training data. Despite the investigation and development of numerous model selection strategies in prior work, the process remains time-consuming, especially given the ever-increasing scale of model repositories. In this paper, we propose a two-phase (coarse-recall and fine-selection) model selection framework, aiming to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets. Specifically, the coarse-recall phase clusters models showcasing similar training performances on benchmark datasets in an offline manner. A light-weight proxy score is subsequently computed between this model cluster and the target dataset, which serves to recall a significantly smaller subset of potential candidate models in a swift manner. In the following fine-selection phase, the final model is chosen by fine-tuning the recalled models on the target dataset with successive halving. To accelerate the process, the final fine-tuning performance of each potential model is predicted by mining the model's convergence trend on the benchmark datasets, which aids in filtering lower performance models more earlier during fine-tuning. Through extensive experimentation on tasks covering natural language processing and computer vision, it has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods. Our code is available at https://github.com/plasware/two-phase-selection.</abstract><doi>10.48550/arxiv.2404.00069</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2404.00069
ispartof
issn
language eng
recordid cdi_arxiv_primary_2404_00069
source arXiv.org
subjects Computer Science - Learning
title A Two-Phase Recall-and-Select Framework for Fast Model Selection
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-12T12%3A49%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Two-Phase%20Recall-and-Select%20Framework%20for%20Fast%20Model%20Selection&rft.au=Cui,%20Jianwei&rft.date=2024-03-28&rft_id=info:doi/10.48550/arxiv.2404.00069&rft_dat=%3Carxiv_GOX%3E2404_00069%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true