Distribution-Level Memory Recall for Continual Learning: Preserving Knowledge and Avoiding Confusion
Continual Learning (CL) aims to enable Deep Neural Networks (DNNs) to learn new data without forgetting previously learned knowledge. The key to achieving this goal is to avoid confusion at the feature level, i.e., avoiding confusion within old tasks and between new and old tasks. Previous prototype...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2024-08 |
---|---|
Hauptverfasser: | , , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Cheng, Shaoxu Geng, Kanglei He, Chiyuan Qiu, Zihuan Xu, Linfeng Qiu, Heqian Wang, Lanxiao Wu, Qingbo Meng, Fanman Li, Hongliang |
description | Continual Learning (CL) aims to enable Deep Neural Networks (DNNs) to learn new data without forgetting previously learned knowledge. The key to achieving this goal is to avoid confusion at the feature level, i.e., avoiding confusion within old tasks and between new and old tasks. Previous prototype-based CL methods generate pseudo features for old knowledge replay by adding Gaussian noise to the centroids of old classes. However, the distribution in the feature space exhibits anisotropy during the incremental process, which prevents the pseudo features from faithfully reproducing the distribution of old knowledge in the feature space, leading to confusion in classification boundaries within old tasks. To address this issue, we propose the Distribution-Level Memory Recall (DMR) method, which uses a Gaussian mixture model to precisely fit the feature distribution of old knowledge at the distribution level and generate pseudo features in the next stage. Furthermore, resistance to confusion at the distribution level is also crucial for multimodal learning, as the problem of multimodal imbalance results in significant differences in feature responses between different modalities, exacerbating confusion within old tasks in prototype-based CL methods. Therefore, we mitigate the multi-modal imbalance problem by using the Inter-modal Guidance and Intra-modal Mining (IGIM) method to guide weaker modalities with prior information from dominant modalities and further explore useful information within modalities. For the second key, We propose the Confusion Index to quantitatively describe a model's ability to distinguish between new and old tasks, and we use the Incremental Mixup Feature Enhancement (IMFE) method to enhance pseudo features with new sample features, alleviating classification confusion between new and old knowledge. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3090036153</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3090036153</sourcerecordid><originalsourceid>FETCH-proquest_journals_30900361533</originalsourceid><addsrcrecordid>eNqNjE0KwjAUhIMgWLR3eOC6kCa2_uykKoIVRNyXaF9LSkw0aSre3goewNUM883MgASM8zhazBgbkdC5hlLK0jlLEh6QciNda-XVt9LoKMcOFRzxbuwbzngTSkFlLGRGt1J7oSBHYbXU9QpOFh3arvdw0OalsKwRhC5h3RlZfuN-VXnX_07IsBLKYfjTMZnutpdsHz2seXp0bdEYb3WPCk6XlPI0Tjj_r_UBR_9GkQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3090036153</pqid></control><display><type>article</type><title>Distribution-Level Memory Recall for Continual Learning: Preserving Knowledge and Avoiding Confusion</title><source>Free E- Journals</source><creator>Cheng, Shaoxu ; Geng, Kanglei ; He, Chiyuan ; Qiu, Zihuan ; Xu, Linfeng ; Qiu, Heqian ; Wang, Lanxiao ; Wu, Qingbo ; Meng, Fanman ; Li, Hongliang</creator><creatorcontrib>Cheng, Shaoxu ; Geng, Kanglei ; He, Chiyuan ; Qiu, Zihuan ; Xu, Linfeng ; Qiu, Heqian ; Wang, Lanxiao ; Wu, Qingbo ; Meng, Fanman ; Li, Hongliang</creatorcontrib><description>Continual Learning (CL) aims to enable Deep Neural Networks (DNNs) to learn new data without forgetting previously learned knowledge. The key to achieving this goal is to avoid confusion at the feature level, i.e., avoiding confusion within old tasks and between new and old tasks. Previous prototype-based CL methods generate pseudo features for old knowledge replay by adding Gaussian noise to the centroids of old classes. However, the distribution in the feature space exhibits anisotropy during the incremental process, which prevents the pseudo features from faithfully reproducing the distribution of old knowledge in the feature space, leading to confusion in classification boundaries within old tasks. To address this issue, we propose the Distribution-Level Memory Recall (DMR) method, which uses a Gaussian mixture model to precisely fit the feature distribution of old knowledge at the distribution level and generate pseudo features in the next stage. Furthermore, resistance to confusion at the distribution level is also crucial for multimodal learning, as the problem of multimodal imbalance results in significant differences in feature responses between different modalities, exacerbating confusion within old tasks in prototype-based CL methods. Therefore, we mitigate the multi-modal imbalance problem by using the Inter-modal Guidance and Intra-modal Mining (IGIM) method to guide weaker modalities with prior information from dominant modalities and further explore useful information within modalities. For the second key, We propose the Confusion Index to quantitatively describe a model's ability to distinguish between new and old tasks, and we use the Incremental Mixup Feature Enhancement (IMFE) method to enhance pseudo features with new sample features, alleviating classification confusion between new and old knowledge.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Anisotropy ; Artificial neural networks ; Centroids ; Classification ; Gaussian process ; Knowledge ; Machine learning ; Memory tasks ; Noise generation ; Normal distribution ; Probabilistic models ; Prototypes ; Random noise ; Recall</subject><ispartof>arXiv.org, 2024-08</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Cheng, Shaoxu</creatorcontrib><creatorcontrib>Geng, Kanglei</creatorcontrib><creatorcontrib>He, Chiyuan</creatorcontrib><creatorcontrib>Qiu, Zihuan</creatorcontrib><creatorcontrib>Xu, Linfeng</creatorcontrib><creatorcontrib>Qiu, Heqian</creatorcontrib><creatorcontrib>Wang, Lanxiao</creatorcontrib><creatorcontrib>Wu, Qingbo</creatorcontrib><creatorcontrib>Meng, Fanman</creatorcontrib><creatorcontrib>Li, Hongliang</creatorcontrib><title>Distribution-Level Memory Recall for Continual Learning: Preserving Knowledge and Avoiding Confusion</title><title>arXiv.org</title><description>Continual Learning (CL) aims to enable Deep Neural Networks (DNNs) to learn new data without forgetting previously learned knowledge. The key to achieving this goal is to avoid confusion at the feature level, i.e., avoiding confusion within old tasks and between new and old tasks. Previous prototype-based CL methods generate pseudo features for old knowledge replay by adding Gaussian noise to the centroids of old classes. However, the distribution in the feature space exhibits anisotropy during the incremental process, which prevents the pseudo features from faithfully reproducing the distribution of old knowledge in the feature space, leading to confusion in classification boundaries within old tasks. To address this issue, we propose the Distribution-Level Memory Recall (DMR) method, which uses a Gaussian mixture model to precisely fit the feature distribution of old knowledge at the distribution level and generate pseudo features in the next stage. Furthermore, resistance to confusion at the distribution level is also crucial for multimodal learning, as the problem of multimodal imbalance results in significant differences in feature responses between different modalities, exacerbating confusion within old tasks in prototype-based CL methods. Therefore, we mitigate the multi-modal imbalance problem by using the Inter-modal Guidance and Intra-modal Mining (IGIM) method to guide weaker modalities with prior information from dominant modalities and further explore useful information within modalities. For the second key, We propose the Confusion Index to quantitatively describe a model's ability to distinguish between new and old tasks, and we use the Incremental Mixup Feature Enhancement (IMFE) method to enhance pseudo features with new sample features, alleviating classification confusion between new and old knowledge.</description><subject>Anisotropy</subject><subject>Artificial neural networks</subject><subject>Centroids</subject><subject>Classification</subject><subject>Gaussian process</subject><subject>Knowledge</subject><subject>Machine learning</subject><subject>Memory tasks</subject><subject>Noise generation</subject><subject>Normal distribution</subject><subject>Probabilistic models</subject><subject>Prototypes</subject><subject>Random noise</subject><subject>Recall</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNqNjE0KwjAUhIMgWLR3eOC6kCa2_uykKoIVRNyXaF9LSkw0aSre3goewNUM883MgASM8zhazBgbkdC5hlLK0jlLEh6QciNda-XVt9LoKMcOFRzxbuwbzngTSkFlLGRGt1J7oSBHYbXU9QpOFh3arvdw0OalsKwRhC5h3RlZfuN-VXnX_07IsBLKYfjTMZnutpdsHz2seXp0bdEYb3WPCk6XlPI0Tjj_r_UBR_9GkQ</recordid><startdate>20240804</startdate><enddate>20240804</enddate><creator>Cheng, Shaoxu</creator><creator>Geng, Kanglei</creator><creator>He, Chiyuan</creator><creator>Qiu, Zihuan</creator><creator>Xu, Linfeng</creator><creator>Qiu, Heqian</creator><creator>Wang, Lanxiao</creator><creator>Wu, Qingbo</creator><creator>Meng, Fanman</creator><creator>Li, Hongliang</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240804</creationdate><title>Distribution-Level Memory Recall for Continual Learning: Preserving Knowledge and Avoiding Confusion</title><author>Cheng, Shaoxu ; Geng, Kanglei ; He, Chiyuan ; Qiu, Zihuan ; Xu, Linfeng ; Qiu, Heqian ; Wang, Lanxiao ; Wu, Qingbo ; Meng, Fanman ; Li, Hongliang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_30900361533</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Anisotropy</topic><topic>Artificial neural networks</topic><topic>Centroids</topic><topic>Classification</topic><topic>Gaussian process</topic><topic>Knowledge</topic><topic>Machine learning</topic><topic>Memory tasks</topic><topic>Noise generation</topic><topic>Normal distribution</topic><topic>Probabilistic models</topic><topic>Prototypes</topic><topic>Random noise</topic><topic>Recall</topic><toplevel>online_resources</toplevel><creatorcontrib>Cheng, Shaoxu</creatorcontrib><creatorcontrib>Geng, Kanglei</creatorcontrib><creatorcontrib>He, Chiyuan</creatorcontrib><creatorcontrib>Qiu, Zihuan</creatorcontrib><creatorcontrib>Xu, Linfeng</creatorcontrib><creatorcontrib>Qiu, Heqian</creatorcontrib><creatorcontrib>Wang, Lanxiao</creatorcontrib><creatorcontrib>Wu, Qingbo</creatorcontrib><creatorcontrib>Meng, Fanman</creatorcontrib><creatorcontrib>Li, Hongliang</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content (ProQuest)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Cheng, Shaoxu</au><au>Geng, Kanglei</au><au>He, Chiyuan</au><au>Qiu, Zihuan</au><au>Xu, Linfeng</au><au>Qiu, Heqian</au><au>Wang, Lanxiao</au><au>Wu, Qingbo</au><au>Meng, Fanman</au><au>Li, Hongliang</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Distribution-Level Memory Recall for Continual Learning: Preserving Knowledge and Avoiding Confusion</atitle><jtitle>arXiv.org</jtitle><date>2024-08-04</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Continual Learning (CL) aims to enable Deep Neural Networks (DNNs) to learn new data without forgetting previously learned knowledge. The key to achieving this goal is to avoid confusion at the feature level, i.e., avoiding confusion within old tasks and between new and old tasks. Previous prototype-based CL methods generate pseudo features for old knowledge replay by adding Gaussian noise to the centroids of old classes. However, the distribution in the feature space exhibits anisotropy during the incremental process, which prevents the pseudo features from faithfully reproducing the distribution of old knowledge in the feature space, leading to confusion in classification boundaries within old tasks. To address this issue, we propose the Distribution-Level Memory Recall (DMR) method, which uses a Gaussian mixture model to precisely fit the feature distribution of old knowledge at the distribution level and generate pseudo features in the next stage. Furthermore, resistance to confusion at the distribution level is also crucial for multimodal learning, as the problem of multimodal imbalance results in significant differences in feature responses between different modalities, exacerbating confusion within old tasks in prototype-based CL methods. Therefore, we mitigate the multi-modal imbalance problem by using the Inter-modal Guidance and Intra-modal Mining (IGIM) method to guide weaker modalities with prior information from dominant modalities and further explore useful information within modalities. For the second key, We propose the Confusion Index to quantitatively describe a model's ability to distinguish between new and old tasks, and we use the Incremental Mixup Feature Enhancement (IMFE) method to enhance pseudo features with new sample features, alleviating classification confusion between new and old knowledge.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-08 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_3090036153 |
source | Free E- Journals |
subjects | Anisotropy Artificial neural networks Centroids Classification Gaussian process Knowledge Machine learning Memory tasks Noise generation Normal distribution Probabilistic models Prototypes Random noise Recall |
title | Distribution-Level Memory Recall for Continual Learning: Preserving Knowledge and Avoiding Confusion |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T17%3A52%3A57IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Distribution-Level%20Memory%20Recall%20for%20Continual%20Learning:%20Preserving%20Knowledge%20and%20Avoiding%20Confusion&rft.jtitle=arXiv.org&rft.au=Cheng,%20Shaoxu&rft.date=2024-08-04&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3090036153%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3090036153&rft_id=info:pmid/&rfr_iscdi=true |