Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token

Deep learning models generally display catastrophic forgetting when learning new data continuously. Many incremental learning approaches address this problem by reusing data from previous tasks while learning new tasks. However, the direct access to past data generates privacy and security concerns....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Choi, Young Jo, Yoo, Min Kyoon, Park, Yu Rang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Choi, Young Jo
Yoo, Min Kyoon
Park, Yu Rang
description Deep learning models generally display catastrophic forgetting when learning new data continuously. Many incremental learning approaches address this problem by reusing data from previous tasks while learning new tasks. However, the direct access to past data generates privacy and security concerns. To address these issues, we present a novel method that preserves previous knowledge without storing previous data. This method is inspired by the architecture of a vision transformer and employs a unique token capable of encapsulating the compressed knowledge of each task. This approach generates task-specific embeddings by directing attention differently based on the task associated with the data, thereby effectively mimicking the impact of having multiple models through tokens. Our method incorporates a distillation process that ensures efficient interactions even after multiple additional learning steps, thereby optimizing the model against forgetting. We measured the performance of our model in terms of accuracy and backward transfer using a benchmark dataset for different task-incremental learning scenarios. Our results demonstrate the superiority of our approach, which achieved the highest accuracy and lowest backward transfer among the compared methods. In addition to presenting a new model, our approach lays the foundation for various extensions within the spectrum of vision-transformer architectures.
doi_str_mv 10.48550/arxiv.2411.05846
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2411_05846</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2411_05846</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2411_058463</originalsourceid><addsrcrecordid>eNqFjr0OgjAUhbs4GPUBnOwLgKBg2I3G2biTWi-0gbbk9vrD29sSd6eTnPPl5GNsnWdpUZVlthX40a90V-R5mpVVcZgzusLjKbVtuRQkPKEblJa8cdgCUexdw7WVCAYsiZ73INDGXltOCri4e7ASIoagwugDZMA4HPlbk-LhtUv8AFI34ZhcB3bJZo3oPax-uWCb8-l2vCSTXz2gNgLHOnrWk-f-P_EFF21K_A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</title><source>arXiv.org</source><creator>Choi, Young Jo ; Yoo, Min Kyoon ; Park, Yu Rang</creator><creatorcontrib>Choi, Young Jo ; Yoo, Min Kyoon ; Park, Yu Rang</creatorcontrib><description>Deep learning models generally display catastrophic forgetting when learning new data continuously. Many incremental learning approaches address this problem by reusing data from previous tasks while learning new tasks. However, the direct access to past data generates privacy and security concerns. To address these issues, we present a novel method that preserves previous knowledge without storing previous data. This method is inspired by the architecture of a vision transformer and employs a unique token capable of encapsulating the compressed knowledge of each task. This approach generates task-specific embeddings by directing attention differently based on the task associated with the data, thereby effectively mimicking the impact of having multiple models through tokens. Our method incorporates a distillation process that ensures efficient interactions even after multiple additional learning steps, thereby optimizing the model against forgetting. We measured the performance of our model in terms of accuracy and backward transfer using a benchmark dataset for different task-incremental learning scenarios. Our results demonstrate the superiority of our approach, which achieved the highest accuracy and lowest backward transfer among the compared methods. In addition to presenting a new model, our approach lays the foundation for various extensions within the spectrum of vision-transformer architectures.</description><identifier>DOI: 10.48550/arxiv.2411.05846</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2024-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2411.05846$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2411.05846$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Choi, Young Jo</creatorcontrib><creatorcontrib>Yoo, Min Kyoon</creatorcontrib><creatorcontrib>Park, Yu Rang</creatorcontrib><title>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</title><description>Deep learning models generally display catastrophic forgetting when learning new data continuously. Many incremental learning approaches address this problem by reusing data from previous tasks while learning new tasks. However, the direct access to past data generates privacy and security concerns. To address these issues, we present a novel method that preserves previous knowledge without storing previous data. This method is inspired by the architecture of a vision transformer and employs a unique token capable of encapsulating the compressed knowledge of each task. This approach generates task-specific embeddings by directing attention differently based on the task associated with the data, thereby effectively mimicking the impact of having multiple models through tokens. Our method incorporates a distillation process that ensures efficient interactions even after multiple additional learning steps, thereby optimizing the model against forgetting. We measured the performance of our model in terms of accuracy and backward transfer using a benchmark dataset for different task-incremental learning scenarios. Our results demonstrate the superiority of our approach, which achieved the highest accuracy and lowest backward transfer among the compared methods. In addition to presenting a new model, our approach lays the foundation for various extensions within the spectrum of vision-transformer architectures.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFjr0OgjAUhbs4GPUBnOwLgKBg2I3G2biTWi-0gbbk9vrD29sSd6eTnPPl5GNsnWdpUZVlthX40a90V-R5mpVVcZgzusLjKbVtuRQkPKEblJa8cdgCUexdw7WVCAYsiZ73INDGXltOCri4e7ASIoagwugDZMA4HPlbk-LhtUv8AFI34ZhcB3bJZo3oPax-uWCb8-l2vCSTXz2gNgLHOnrWk-f-P_EFF21K_A</recordid><startdate>20241106</startdate><enddate>20241106</enddate><creator>Choi, Young Jo</creator><creator>Yoo, Min Kyoon</creator><creator>Park, Yu Rang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241106</creationdate><title>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</title><author>Choi, Young Jo ; Yoo, Min Kyoon ; Park, Yu Rang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2411_058463</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Choi, Young Jo</creatorcontrib><creatorcontrib>Yoo, Min Kyoon</creatorcontrib><creatorcontrib>Park, Yu Rang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Choi, Young Jo</au><au>Yoo, Min Kyoon</au><au>Park, Yu Rang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token</atitle><date>2024-11-06</date><risdate>2024</risdate><abstract>Deep learning models generally display catastrophic forgetting when learning new data continuously. Many incremental learning approaches address this problem by reusing data from previous tasks while learning new tasks. However, the direct access to past data generates privacy and security concerns. To address these issues, we present a novel method that preserves previous knowledge without storing previous data. This method is inspired by the architecture of a vision transformer and employs a unique token capable of encapsulating the compressed knowledge of each task. This approach generates task-specific embeddings by directing attention differently based on the task associated with the data, thereby effectively mimicking the impact of having multiple models through tokens. Our method incorporates a distillation process that ensures efficient interactions even after multiple additional learning steps, thereby optimizing the model against forgetting. We measured the performance of our model in terms of accuracy and backward transfer using a benchmark dataset for different task-incremental learning scenarios. Our results demonstrate the superiority of our approach, which achieved the highest accuracy and lowest backward transfer among the compared methods. In addition to presenting a new model, our approach lays the foundation for various extensions within the spectrum of vision-transformer architectures.</abstract><doi>10.48550/arxiv.2411.05846</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2411.05846
ispartof
issn
language eng
recordid cdi_arxiv_primary_2411_05846
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
title Reducing catastrophic forgetting of incremental learning in the absence of rehearsal memory with task-specific token
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-29T14%3A13%3A06IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Reducing%20catastrophic%20forgetting%20of%20incremental%20learning%20in%20the%20absence%20of%20rehearsal%20memory%20with%20task-specific%20token&rft.au=Choi,%20Young%20Jo&rft.date=2024-11-06&rft_id=info:doi/10.48550/arxiv.2411.05846&rft_dat=%3Carxiv_GOX%3E2411_05846%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true