Adaptive Transfer Learning: a simple but effective transfer learning
Transfer learning (TL) leverages previously obtained knowledge to learn new tasks efficiently and has been used to train deep learning (DL) models with limited amount of data. When TL is applied to DL, pretrained (teacher) models are fine-tuned to build domain specific (student) models. This fine-tu...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Lee, Jung H Kvinge, Henry J Howland, Scott New, Zachary Buckheit, John Phillips, Lauren A Skomski, Elliott Hibler, Jessica Corley, Courtney D Hodas, Nathan O |
description | Transfer learning (TL) leverages previously obtained knowledge to learn new
tasks efficiently and has been used to train deep learning (DL) models with
limited amount of data. When TL is applied to DL, pretrained (teacher) models
are fine-tuned to build domain specific (student) models. This fine-tuning
relies on the fact that DL model can be decomposed to classifiers and feature
extractors, and a line of studies showed that the same feature extractors can
be used to train classifiers on multiple tasks. Furthermore, recent studies
proposed multiple algorithms that can fine-tune teacher models' feature
extractors to train student models more efficiently. We note that regardless of
the fine-tuning of feature extractors, the classifiers of student models are
trained with final outputs of feature extractors (i.e., the outputs of
penultimate layers). However, a recent study suggested that feature maps in
ResNets across layers could be functionally equivalent, raising the possibility
that feature maps inside the feature extractors can also be used to train
student models' classifiers. Inspired by this study, we tested if feature maps
in the hidden layers of the teacher models can be used to improve the student
models' accuracy (i.e., TL's efficiency). Specifically, we developed 'adaptive
transfer learning (ATL)', which can choose an optimal set of feature maps for
TL, and tested it in the few-shot learning setting. Our empirical evaluations
suggest that ATL can help DL models learn more efficiently, especially when
available examples are limited. |
doi_str_mv | 10.48550/arxiv.2111.10937 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2111_10937</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2111_10937</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-f2384882c65e708e977b5407529165697155ef9adf5e0943b2fa13267e4a58bf3</originalsourceid><addsrcrecordid>eNo1z8tOwzAQhWFvWKDCA7DCL5Dg23hsdlW5SpG6yT6atGNkKY0iJ1Tw9ohQVmfz6Ui_EHda1S4AqAcqX_lcG611rVW0eC2etkealnxm2RYa58RFNkxlzOPHoyQ559M0sOw_F8kp8WGVy78cLvJGXCUaZr697Ea0L8_t7q1q9q_vu21TkUeskrHBhWAOHhhV4IjYg1MIJmoPPqIG4BTpmIBVdLY3ibQ1HtkRhD7Zjbj_u10zuqnkE5Xv7jenW3PsDwEMREM</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Adaptive Transfer Learning: a simple but effective transfer learning</title><source>arXiv.org</source><creator>Lee, Jung H ; Kvinge, Henry J ; Howland, Scott ; New, Zachary ; Buckheit, John ; Phillips, Lauren A ; Skomski, Elliott ; Hibler, Jessica ; Corley, Courtney D ; Hodas, Nathan O</creator><creatorcontrib>Lee, Jung H ; Kvinge, Henry J ; Howland, Scott ; New, Zachary ; Buckheit, John ; Phillips, Lauren A ; Skomski, Elliott ; Hibler, Jessica ; Corley, Courtney D ; Hodas, Nathan O</creatorcontrib><description>Transfer learning (TL) leverages previously obtained knowledge to learn new
tasks efficiently and has been used to train deep learning (DL) models with
limited amount of data. When TL is applied to DL, pretrained (teacher) models
are fine-tuned to build domain specific (student) models. This fine-tuning
relies on the fact that DL model can be decomposed to classifiers and feature
extractors, and a line of studies showed that the same feature extractors can
be used to train classifiers on multiple tasks. Furthermore, recent studies
proposed multiple algorithms that can fine-tune teacher models' feature
extractors to train student models more efficiently. We note that regardless of
the fine-tuning of feature extractors, the classifiers of student models are
trained with final outputs of feature extractors (i.e., the outputs of
penultimate layers). However, a recent study suggested that feature maps in
ResNets across layers could be functionally equivalent, raising the possibility
that feature maps inside the feature extractors can also be used to train
student models' classifiers. Inspired by this study, we tested if feature maps
in the hidden layers of the teacher models can be used to improve the student
models' accuracy (i.e., TL's efficiency). Specifically, we developed 'adaptive
transfer learning (ATL)', which can choose an optimal set of feature maps for
TL, and tested it in the few-shot learning setting. Our empirical evaluations
suggest that ATL can help DL models learn more efficiently, especially when
available examples are limited.</description><identifier>DOI: 10.48550/arxiv.2111.10937</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2021-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2111.10937$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2111.10937$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Lee, Jung H</creatorcontrib><creatorcontrib>Kvinge, Henry J</creatorcontrib><creatorcontrib>Howland, Scott</creatorcontrib><creatorcontrib>New, Zachary</creatorcontrib><creatorcontrib>Buckheit, John</creatorcontrib><creatorcontrib>Phillips, Lauren A</creatorcontrib><creatorcontrib>Skomski, Elliott</creatorcontrib><creatorcontrib>Hibler, Jessica</creatorcontrib><creatorcontrib>Corley, Courtney D</creatorcontrib><creatorcontrib>Hodas, Nathan O</creatorcontrib><title>Adaptive Transfer Learning: a simple but effective transfer learning</title><description>Transfer learning (TL) leverages previously obtained knowledge to learn new
tasks efficiently and has been used to train deep learning (DL) models with
limited amount of data. When TL is applied to DL, pretrained (teacher) models
are fine-tuned to build domain specific (student) models. This fine-tuning
relies on the fact that DL model can be decomposed to classifiers and feature
extractors, and a line of studies showed that the same feature extractors can
be used to train classifiers on multiple tasks. Furthermore, recent studies
proposed multiple algorithms that can fine-tune teacher models' feature
extractors to train student models more efficiently. We note that regardless of
the fine-tuning of feature extractors, the classifiers of student models are
trained with final outputs of feature extractors (i.e., the outputs of
penultimate layers). However, a recent study suggested that feature maps in
ResNets across layers could be functionally equivalent, raising the possibility
that feature maps inside the feature extractors can also be used to train
student models' classifiers. Inspired by this study, we tested if feature maps
in the hidden layers of the teacher models can be used to improve the student
models' accuracy (i.e., TL's efficiency). Specifically, we developed 'adaptive
transfer learning (ATL)', which can choose an optimal set of feature maps for
TL, and tested it in the few-shot learning setting. Our empirical evaluations
suggest that ATL can help DL models learn more efficiently, especially when
available examples are limited.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNo1z8tOwzAQhWFvWKDCA7DCL5Dg23hsdlW5SpG6yT6atGNkKY0iJ1Tw9ohQVmfz6Ui_EHda1S4AqAcqX_lcG611rVW0eC2etkealnxm2RYa58RFNkxlzOPHoyQ559M0sOw_F8kp8WGVy78cLvJGXCUaZr697Ea0L8_t7q1q9q_vu21TkUeskrHBhWAOHhhV4IjYg1MIJmoPPqIG4BTpmIBVdLY3ibQ1HtkRhD7Zjbj_u10zuqnkE5Xv7jenW3PsDwEMREM</recordid><startdate>20211121</startdate><enddate>20211121</enddate><creator>Lee, Jung H</creator><creator>Kvinge, Henry J</creator><creator>Howland, Scott</creator><creator>New, Zachary</creator><creator>Buckheit, John</creator><creator>Phillips, Lauren A</creator><creator>Skomski, Elliott</creator><creator>Hibler, Jessica</creator><creator>Corley, Courtney D</creator><creator>Hodas, Nathan O</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20211121</creationdate><title>Adaptive Transfer Learning: a simple but effective transfer learning</title><author>Lee, Jung H ; Kvinge, Henry J ; Howland, Scott ; New, Zachary ; Buckheit, John ; Phillips, Lauren A ; Skomski, Elliott ; Hibler, Jessica ; Corley, Courtney D ; Hodas, Nathan O</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-f2384882c65e708e977b5407529165697155ef9adf5e0943b2fa13267e4a58bf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Lee, Jung H</creatorcontrib><creatorcontrib>Kvinge, Henry J</creatorcontrib><creatorcontrib>Howland, Scott</creatorcontrib><creatorcontrib>New, Zachary</creatorcontrib><creatorcontrib>Buckheit, John</creatorcontrib><creatorcontrib>Phillips, Lauren A</creatorcontrib><creatorcontrib>Skomski, Elliott</creatorcontrib><creatorcontrib>Hibler, Jessica</creatorcontrib><creatorcontrib>Corley, Courtney D</creatorcontrib><creatorcontrib>Hodas, Nathan O</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Lee, Jung H</au><au>Kvinge, Henry J</au><au>Howland, Scott</au><au>New, Zachary</au><au>Buckheit, John</au><au>Phillips, Lauren A</au><au>Skomski, Elliott</au><au>Hibler, Jessica</au><au>Corley, Courtney D</au><au>Hodas, Nathan O</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Adaptive Transfer Learning: a simple but effective transfer learning</atitle><date>2021-11-21</date><risdate>2021</risdate><abstract>Transfer learning (TL) leverages previously obtained knowledge to learn new
tasks efficiently and has been used to train deep learning (DL) models with
limited amount of data. When TL is applied to DL, pretrained (teacher) models
are fine-tuned to build domain specific (student) models. This fine-tuning
relies on the fact that DL model can be decomposed to classifiers and feature
extractors, and a line of studies showed that the same feature extractors can
be used to train classifiers on multiple tasks. Furthermore, recent studies
proposed multiple algorithms that can fine-tune teacher models' feature
extractors to train student models more efficiently. We note that regardless of
the fine-tuning of feature extractors, the classifiers of student models are
trained with final outputs of feature extractors (i.e., the outputs of
penultimate layers). However, a recent study suggested that feature maps in
ResNets across layers could be functionally equivalent, raising the possibility
that feature maps inside the feature extractors can also be used to train
student models' classifiers. Inspired by this study, we tested if feature maps
in the hidden layers of the teacher models can be used to improve the student
models' accuracy (i.e., TL's efficiency). Specifically, we developed 'adaptive
transfer learning (ATL)', which can choose an optimal set of feature maps for
TL, and tested it in the few-shot learning setting. Our empirical evaluations
suggest that ATL can help DL models learn more efficiently, especially when
available examples are limited.</abstract><doi>10.48550/arxiv.2111.10937</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2111.10937 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2111_10937 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | Adaptive Transfer Learning: a simple but effective transfer learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T11%3A32%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Adaptive%20Transfer%20Learning:%20a%20simple%20but%20effective%20transfer%20learning&rft.au=Lee,%20Jung%20H&rft.date=2021-11-21&rft_id=info:doi/10.48550/arxiv.2111.10937&rft_dat=%3Carxiv_GOX%3E2111_10937%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |