A Two-Phase Recall-and-Select Framework for Fast Model Selection
As the ubiquity of deep learning in various machine learning applications has amplified, a proliferation of neural network models has been trained and shared on public model repositories. In the context of a targeted machine learning assignment, utilizing an apt source model as a starting point typi...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Cui, Jianwei Shi, Wenhang Tao, Honglin Lu, Wei Du, Xiaoyong |
description | As the ubiquity of deep learning in various machine learning applications has
amplified, a proliferation of neural network models has been trained and shared
on public model repositories. In the context of a targeted machine learning
assignment, utilizing an apt source model as a starting point typically
outperforms the strategy of training from scratch, particularly with limited
training data. Despite the investigation and development of numerous model
selection strategies in prior work, the process remains time-consuming,
especially given the ever-increasing scale of model repositories. In this
paper, we propose a two-phase (coarse-recall and fine-selection) model
selection framework, aiming to enhance the efficiency of selecting a robust
model by leveraging the models' training performances on benchmark datasets.
Specifically, the coarse-recall phase clusters models showcasing similar
training performances on benchmark datasets in an offline manner. A
light-weight proxy score is subsequently computed between this model cluster
and the target dataset, which serves to recall a significantly smaller subset
of potential candidate models in a swift manner. In the following
fine-selection phase, the final model is chosen by fine-tuning the recalled
models on the target dataset with successive halving. To accelerate the
process, the final fine-tuning performance of each potential model is predicted
by mining the model's convergence trend on the benchmark datasets, which aids
in filtering lower performance models more earlier during fine-tuning. Through
extensive experimentation on tasks covering natural language processing and
computer vision, it has been demonstrated that the proposed methodology
facilitates the selection of a high-performing model at a rate about 3x times
faster than conventional baseline methods. Our code is available at
https://github.com/plasware/two-phase-selection. |
doi_str_mv | 10.48550/arxiv.2404.00069 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2404_00069</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2404_00069</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-7c1ce280216d255b691a7023c6f885963a0c0fa2e036e90494a21d6a2303668a3</originalsourceid><addsrcrecordid>eNotz81KAzEUBeBsupDWB3BlXiDTm99JdpbiqFBRdPbDNXMHB9OmZIrVt1dbVwfOgQMfY1cSKuOthSWWr_GzUgZMBQAuXLCbFW-PWTy_40T8hSKmJHDXi1dKFA-8KbilYy4ffMiFNzgd-GPuKfHzPubdgs0GTBNd_uectc1tu74Xm6e7h_VqI9DVQdRRRlIelHS9svbNBYk1KB3d4L0NTiNEGFARaEcBTDCoZO9Q6d_CedRzdn2-PRG6fRm3WL67P0p3ougf78hBYA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Two-Phase Recall-and-Select Framework for Fast Model Selection</title><source>arXiv.org</source><creator>Cui, Jianwei ; Shi, Wenhang ; Tao, Honglin ; Lu, Wei ; Du, Xiaoyong</creator><creatorcontrib>Cui, Jianwei ; Shi, Wenhang ; Tao, Honglin ; Lu, Wei ; Du, Xiaoyong</creatorcontrib><description>As the ubiquity of deep learning in various machine learning applications has
amplified, a proliferation of neural network models has been trained and shared
on public model repositories. In the context of a targeted machine learning
assignment, utilizing an apt source model as a starting point typically
outperforms the strategy of training from scratch, particularly with limited
training data. Despite the investigation and development of numerous model
selection strategies in prior work, the process remains time-consuming,
especially given the ever-increasing scale of model repositories. In this
paper, we propose a two-phase (coarse-recall and fine-selection) model
selection framework, aiming to enhance the efficiency of selecting a robust
model by leveraging the models' training performances on benchmark datasets.
Specifically, the coarse-recall phase clusters models showcasing similar
training performances on benchmark datasets in an offline manner. A
light-weight proxy score is subsequently computed between this model cluster
and the target dataset, which serves to recall a significantly smaller subset
of potential candidate models in a swift manner. In the following
fine-selection phase, the final model is chosen by fine-tuning the recalled
models on the target dataset with successive halving. To accelerate the
process, the final fine-tuning performance of each potential model is predicted
by mining the model's convergence trend on the benchmark datasets, which aids
in filtering lower performance models more earlier during fine-tuning. Through
extensive experimentation on tasks covering natural language processing and
computer vision, it has been demonstrated that the proposed methodology
facilitates the selection of a high-performing model at a rate about 3x times
faster than conventional baseline methods. Our code is available at
https://github.com/plasware/two-phase-selection.</description><identifier>DOI: 10.48550/arxiv.2404.00069</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2024-03</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2404.00069$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2404.00069$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Cui, Jianwei</creatorcontrib><creatorcontrib>Shi, Wenhang</creatorcontrib><creatorcontrib>Tao, Honglin</creatorcontrib><creatorcontrib>Lu, Wei</creatorcontrib><creatorcontrib>Du, Xiaoyong</creatorcontrib><title>A Two-Phase Recall-and-Select Framework for Fast Model Selection</title><description>As the ubiquity of deep learning in various machine learning applications has
amplified, a proliferation of neural network models has been trained and shared
on public model repositories. In the context of a targeted machine learning
assignment, utilizing an apt source model as a starting point typically
outperforms the strategy of training from scratch, particularly with limited
training data. Despite the investigation and development of numerous model
selection strategies in prior work, the process remains time-consuming,
especially given the ever-increasing scale of model repositories. In this
paper, we propose a two-phase (coarse-recall and fine-selection) model
selection framework, aiming to enhance the efficiency of selecting a robust
model by leveraging the models' training performances on benchmark datasets.
Specifically, the coarse-recall phase clusters models showcasing similar
training performances on benchmark datasets in an offline manner. A
light-weight proxy score is subsequently computed between this model cluster
and the target dataset, which serves to recall a significantly smaller subset
of potential candidate models in a swift manner. In the following
fine-selection phase, the final model is chosen by fine-tuning the recalled
models on the target dataset with successive halving. To accelerate the
process, the final fine-tuning performance of each potential model is predicted
by mining the model's convergence trend on the benchmark datasets, which aids
in filtering lower performance models more earlier during fine-tuning. Through
extensive experimentation on tasks covering natural language processing and
computer vision, it has been demonstrated that the proposed methodology
facilitates the selection of a high-performing model at a rate about 3x times
faster than conventional baseline methods. Our code is available at
https://github.com/plasware/two-phase-selection.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz81KAzEUBeBsupDWB3BlXiDTm99JdpbiqFBRdPbDNXMHB9OmZIrVt1dbVwfOgQMfY1cSKuOthSWWr_GzUgZMBQAuXLCbFW-PWTy_40T8hSKmJHDXi1dKFA-8KbilYy4ffMiFNzgd-GPuKfHzPubdgs0GTBNd_uectc1tu74Xm6e7h_VqI9DVQdRRRlIelHS9svbNBYk1KB3d4L0NTiNEGFARaEcBTDCoZO9Q6d_CedRzdn2-PRG6fRm3WL67P0p3ougf78hBYA</recordid><startdate>20240328</startdate><enddate>20240328</enddate><creator>Cui, Jianwei</creator><creator>Shi, Wenhang</creator><creator>Tao, Honglin</creator><creator>Lu, Wei</creator><creator>Du, Xiaoyong</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240328</creationdate><title>A Two-Phase Recall-and-Select Framework for Fast Model Selection</title><author>Cui, Jianwei ; Shi, Wenhang ; Tao, Honglin ; Lu, Wei ; Du, Xiaoyong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-7c1ce280216d255b691a7023c6f885963a0c0fa2e036e90494a21d6a2303668a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Cui, Jianwei</creatorcontrib><creatorcontrib>Shi, Wenhang</creatorcontrib><creatorcontrib>Tao, Honglin</creatorcontrib><creatorcontrib>Lu, Wei</creatorcontrib><creatorcontrib>Du, Xiaoyong</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Cui, Jianwei</au><au>Shi, Wenhang</au><au>Tao, Honglin</au><au>Lu, Wei</au><au>Du, Xiaoyong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Two-Phase Recall-and-Select Framework for Fast Model Selection</atitle><date>2024-03-28</date><risdate>2024</risdate><abstract>As the ubiquity of deep learning in various machine learning applications has
amplified, a proliferation of neural network models has been trained and shared
on public model repositories. In the context of a targeted machine learning
assignment, utilizing an apt source model as a starting point typically
outperforms the strategy of training from scratch, particularly with limited
training data. Despite the investigation and development of numerous model
selection strategies in prior work, the process remains time-consuming,
especially given the ever-increasing scale of model repositories. In this
paper, we propose a two-phase (coarse-recall and fine-selection) model
selection framework, aiming to enhance the efficiency of selecting a robust
model by leveraging the models' training performances on benchmark datasets.
Specifically, the coarse-recall phase clusters models showcasing similar
training performances on benchmark datasets in an offline manner. A
light-weight proxy score is subsequently computed between this model cluster
and the target dataset, which serves to recall a significantly smaller subset
of potential candidate models in a swift manner. In the following
fine-selection phase, the final model is chosen by fine-tuning the recalled
models on the target dataset with successive halving. To accelerate the
process, the final fine-tuning performance of each potential model is predicted
by mining the model's convergence trend on the benchmark datasets, which aids
in filtering lower performance models more earlier during fine-tuning. Through
extensive experimentation on tasks covering natural language processing and
computer vision, it has been demonstrated that the proposed methodology
facilitates the selection of a high-performing model at a rate about 3x times
faster than conventional baseline methods. Our code is available at
https://github.com/plasware/two-phase-selection.</abstract><doi>10.48550/arxiv.2404.00069</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2404.00069 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2404_00069 |
source | arXiv.org |
subjects | Computer Science - Learning |
title | A Two-Phase Recall-and-Select Framework for Fast Model Selection |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-12T12%3A49%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Two-Phase%20Recall-and-Select%20Framework%20for%20Fast%20Model%20Selection&rft.au=Cui,%20Jianwei&rft.date=2024-03-28&rft_id=info:doi/10.48550/arxiv.2404.00069&rft_dat=%3Carxiv_GOX%3E2404_00069%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |