Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token
Deep learning models generally display catastrophic forgetting when learning new data continuously. Many incremental learning approaches address this problem by reusing data from previous tasks while learning new tasks. However, the direct access to past data generates privacy and security concerns....
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Choi, Young Jo Yoo, Min Kyoon Park, Yu Rang |
description | Deep learning models generally display catastrophic forgetting when learning
new data continuously. Many incremental learning approaches address this
problem by reusing data from previous tasks while learning new tasks. However,
the direct access to past data generates privacy and security concerns. To
address these issues, we present a novel method that preserves previous
knowledge without storing previous data. This method is inspired by the
architecture of a vision transformer and employs a unique token capable of
encapsulating the compressed knowledge of each task. This approach generates
task-specific embeddings by directing attention differently based on the task
associated with the data, thereby effectively mimicking the impact of having
multiple models through tokens. Our method incorporates a distillation process
that ensures efficient interactions even after multiple additional learning
steps, thereby optimizing the model against forgetting. We measured the
performance of our model in terms of accuracy and backward transfer using a
benchmark dataset for different task-incremental learning scenarios. Our
results demonstrate the superiority of our approach, which achieved the highest
accuracy and lowest backward transfer among the compared methods. In addition
to presenting a new model, our approach lays the foundation for various
extensions within the spectrum of vision-transformer architectures. |
doi_str_mv | 10.48550/arxiv.2411.05846 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2411_05846</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2411_05846</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2411_058463</originalsourceid><addsrcrecordid>eNqFjr0OgjAUhbs4GPUBnOwLgKBg2I3G2biTWi-0gbbk9vrD29sSd6eTnPPl5GNsnWdpUZVlthX40a90V-R5mpVVcZgzusLjKbVtuRQkPKEblJa8cdgCUexdw7WVCAYsiZ73INDGXltOCri4e7ASIoagwugDZMA4HPlbk-LhtUv8AFI34ZhcB3bJZo3oPax-uWCb8-l2vCSTXz2gNgLHOnrWk-f-P_EFF21K_A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</title><source>arXiv.org</source><creator>Choi, Young Jo ; Yoo, Min Kyoon ; Park, Yu Rang</creator><creatorcontrib>Choi, Young Jo ; Yoo, Min Kyoon ; Park, Yu Rang</creatorcontrib><description>Deep learning models generally display catastrophic forgetting when learning
new data continuously. Many incremental learning approaches address this
problem by reusing data from previous tasks while learning new tasks. However,
the direct access to past data generates privacy and security concerns. To
address these issues, we present a novel method that preserves previous
knowledge without storing previous data. This method is inspired by the
architecture of a vision transformer and employs a unique token capable of
encapsulating the compressed knowledge of each task. This approach generates
task-specific embeddings by directing attention differently based on the task
associated with the data, thereby effectively mimicking the impact of having
multiple models through tokens. Our method incorporates a distillation process
that ensures efficient interactions even after multiple additional learning
steps, thereby optimizing the model against forgetting. We measured the
performance of our model in terms of accuracy and backward transfer using a
benchmark dataset for different task-incremental learning scenarios. Our
results demonstrate the superiority of our approach, which achieved the highest
accuracy and lowest backward transfer among the compared methods. In addition
to presenting a new model, our approach lays the foundation for various
extensions within the spectrum of vision-transformer architectures.</description><identifier>DOI: 10.48550/arxiv.2411.05846</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2024-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2411.05846$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2411.05846$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Choi, Young Jo</creatorcontrib><creatorcontrib>Yoo, Min Kyoon</creatorcontrib><creatorcontrib>Park, Yu Rang</creatorcontrib><title>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</title><description>Deep learning models generally display catastrophic forgetting when learning
new data continuously. Many incremental learning approaches address this
problem by reusing data from previous tasks while learning new tasks. However,
the direct access to past data generates privacy and security concerns. To
address these issues, we present a novel method that preserves previous
knowledge without storing previous data. This method is inspired by the
architecture of a vision transformer and employs a unique token capable of
encapsulating the compressed knowledge of each task. This approach generates
task-specific embeddings by directing attention differently based on the task
associated with the data, thereby effectively mimicking the impact of having
multiple models through tokens. Our method incorporates a distillation process
that ensures efficient interactions even after multiple additional learning
steps, thereby optimizing the model against forgetting. We measured the
performance of our model in terms of accuracy and backward transfer using a
benchmark dataset for different task-incremental learning scenarios. Our
results demonstrate the superiority of our approach, which achieved the highest
accuracy and lowest backward transfer among the compared methods. In addition
to presenting a new model, our approach lays the foundation for various
extensions within the spectrum of vision-transformer architectures.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFjr0OgjAUhbs4GPUBnOwLgKBg2I3G2biTWi-0gbbk9vrD29sSd6eTnPPl5GNsnWdpUZVlthX40a90V-R5mpVVcZgzusLjKbVtuRQkPKEblJa8cdgCUexdw7WVCAYsiZ73INDGXltOCri4e7ASIoagwugDZMA4HPlbk-LhtUv8AFI34ZhcB3bJZo3oPax-uWCb8-l2vCSTXz2gNgLHOnrWk-f-P_EFF21K_A</recordid><startdate>20241106</startdate><enddate>20241106</enddate><creator>Choi, Young Jo</creator><creator>Yoo, Min Kyoon</creator><creator>Park, Yu Rang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241106</creationdate><title>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</title><author>Choi, Young Jo ; Yoo, Min Kyoon ; Park, Yu Rang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2411_058463</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Choi, Young Jo</creatorcontrib><creatorcontrib>Yoo, Min Kyoon</creatorcontrib><creatorcontrib>Park, Yu Rang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Choi, Young Jo</au><au>Yoo, Min Kyoon</au><au>Park, Yu Rang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</atitle><date>2024-11-06</date><risdate>2024</risdate><abstract>Deep learning models generally display catastrophic forgetting when learning
new data continuously. Many incremental learning approaches address this
problem by reusing data from previous tasks while learning new tasks. However,
the direct access to past data generates privacy and security concerns. To
address these issues, we present a novel method that preserves previous
knowledge without storing previous data. This method is inspired by the
architecture of a vision transformer and employs a unique token capable of
encapsulating the compressed knowledge of each task. This approach generates
task-specific embeddings by directing attention differently based on the task
associated with the data, thereby effectively mimicking the impact of having
multiple models through tokens. Our method incorporates a distillation process
that ensures efficient interactions even after multiple additional learning
steps, thereby optimizing the model against forgetting. We measured the
performance of our model in terms of accuracy and backward transfer using a
benchmark dataset for different task-incremental learning scenarios. Our
results demonstrate the superiority of our approach, which achieved the highest
accuracy and lowest backward transfer among the compared methods. In addition
to presenting a new model, our approach lays the foundation for various
extensions within the spectrum of vision-transformer architectures.</abstract><doi>10.48550/arxiv.2411.05846</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2411.05846 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2411_05846 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-29T14%3A13%3A06IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Reducing%20catastrophic%20forgetting%20of%20incremental%20learning%20in%20the%20absence%20of%20rehearsal%20memory%20with%20task-specific%20token&rft.au=Choi,%20Young%20Jo&rft.date=2024-11-06&rft_id=info:doi/10.48550/arxiv.2411.05846&rft_dat=%3Carxiv_GOX%3E2411_05846%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |