Neuromimetic metaplasticity for adaptive continual learning
Conventional intelligent systems based on deep neural network (DNN) models encounter challenges in achieving human-like continual learning due to catastrophic forgetting. Here, we propose a metaplasticity model inspired by human working memory, enabling DNNs to perform catastrophic forgetting-free c...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Cho, Suhee Lee, Hyeonsu Baek, Seungdae Paik, Se-Bum |
description | Conventional intelligent systems based on deep neural network (DNN) models
encounter challenges in achieving human-like continual learning due to
catastrophic forgetting. Here, we propose a metaplasticity model inspired by
human working memory, enabling DNNs to perform catastrophic forgetting-free
continual learning without any pre- or post-processing. A key aspect of our
approach involves implementing distinct types of synapses from stable to
flexible, and randomly intermixing them to train synaptic connections with
different degrees of flexibility. This strategy allowed the network to
successfully learn a continuous stream of information, even under unexpected
changes in input length. The model achieved a balanced tradeoff between memory
capacity and performance without requiring additional training or structural
modifications, dynamically allocating memory resources to retain both old and
new information. Furthermore, the model demonstrated robustness against data
poisoning attacks by selectively filtering out erroneous memories, leveraging
the Hebb repetition effect to reinforce the retention of significant data. |
doi_str_mv | 10.48550/arxiv.2407.07133 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2407_07133</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2407_07133</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2407_071333</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zMwNzQ25mSw9kstLcrPzcxNLclMVgCSiQU5icVAdmZJpUJafpFCYkpiQUlmWapCcn5eSWZeaWKOQk5qYlFeZl46DwNrWmJOcSovlOZmkHdzDXH20AVbE19QlJmbWFQZD7IuHmydMWEVAEcpNfA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Neuromimetic metaplasticity for adaptive continual learning</title><source>arXiv.org</source><creator>Cho, Suhee ; Lee, Hyeonsu ; Baek, Seungdae ; Paik, Se-Bum</creator><creatorcontrib>Cho, Suhee ; Lee, Hyeonsu ; Baek, Seungdae ; Paik, Se-Bum</creatorcontrib><description>Conventional intelligent systems based on deep neural network (DNN) models
encounter challenges in achieving human-like continual learning due to
catastrophic forgetting. Here, we propose a metaplasticity model inspired by
human working memory, enabling DNNs to perform catastrophic forgetting-free
continual learning without any pre- or post-processing. A key aspect of our
approach involves implementing distinct types of synapses from stable to
flexible, and randomly intermixing them to train synaptic connections with
different degrees of flexibility. This strategy allowed the network to
successfully learn a continuous stream of information, even under unexpected
changes in input length. The model achieved a balanced tradeoff between memory
capacity and performance without requiring additional training or structural
modifications, dynamically allocating memory resources to retain both old and
new information. Furthermore, the model demonstrated robustness against data
poisoning attacks by selectively filtering out erroneous memories, leveraging
the Hebb repetition effect to reinforce the retention of significant data.</description><identifier>DOI: 10.48550/arxiv.2407.07133</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning ; Computer Science - Neural and Evolutionary Computing</subject><creationdate>2024-07</creationdate><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2407.07133$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2407.07133$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Cho, Suhee</creatorcontrib><creatorcontrib>Lee, Hyeonsu</creatorcontrib><creatorcontrib>Baek, Seungdae</creatorcontrib><creatorcontrib>Paik, Se-Bum</creatorcontrib><title>Neuromimetic metaplasticity for adaptive continual learning</title><description>Conventional intelligent systems based on deep neural network (DNN) models
encounter challenges in achieving human-like continual learning due to
catastrophic forgetting. Here, we propose a metaplasticity model inspired by
human working memory, enabling DNNs to perform catastrophic forgetting-free
continual learning without any pre- or post-processing. A key aspect of our
approach involves implementing distinct types of synapses from stable to
flexible, and randomly intermixing them to train synaptic connections with
different degrees of flexibility. This strategy allowed the network to
successfully learn a continuous stream of information, even under unexpected
changes in input length. The model achieved a balanced tradeoff between memory
capacity and performance without requiring additional training or structural
modifications, dynamically allocating memory resources to retain both old and
new information. Furthermore, the model demonstrated robustness against data
poisoning attacks by selectively filtering out erroneous memories, leveraging
the Hebb repetition effect to reinforce the retention of significant data.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Neural and Evolutionary Computing</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zMwNzQ25mSw9kstLcrPzcxNLclMVgCSiQU5icVAdmZJpUJafpFCYkpiQUlmWapCcn5eSWZeaWKOQk5qYlFeZl46DwNrWmJOcSovlOZmkHdzDXH20AVbE19QlJmbWFQZD7IuHmydMWEVAEcpNfA</recordid><startdate>20240709</startdate><enddate>20240709</enddate><creator>Cho, Suhee</creator><creator>Lee, Hyeonsu</creator><creator>Baek, Seungdae</creator><creator>Paik, Se-Bum</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240709</creationdate><title>Neuromimetic metaplasticity for adaptive continual learning</title><author>Cho, Suhee ; Lee, Hyeonsu ; Baek, Seungdae ; Paik, Se-Bum</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2407_071333</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Neural and Evolutionary Computing</topic><toplevel>online_resources</toplevel><creatorcontrib>Cho, Suhee</creatorcontrib><creatorcontrib>Lee, Hyeonsu</creatorcontrib><creatorcontrib>Baek, Seungdae</creatorcontrib><creatorcontrib>Paik, Se-Bum</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Cho, Suhee</au><au>Lee, Hyeonsu</au><au>Baek, Seungdae</au><au>Paik, Se-Bum</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Neuromimetic metaplasticity for adaptive continual learning</atitle><date>2024-07-09</date><risdate>2024</risdate><abstract>Conventional intelligent systems based on deep neural network (DNN) models
encounter challenges in achieving human-like continual learning due to
catastrophic forgetting. Here, we propose a metaplasticity model inspired by
human working memory, enabling DNNs to perform catastrophic forgetting-free
continual learning without any pre- or post-processing. A key aspect of our
approach involves implementing distinct types of synapses from stable to
flexible, and randomly intermixing them to train synaptic connections with
different degrees of flexibility. This strategy allowed the network to
successfully learn a continuous stream of information, even under unexpected
changes in input length. The model achieved a balanced tradeoff between memory
capacity and performance without requiring additional training or structural
modifications, dynamically allocating memory resources to retain both old and
new information. Furthermore, the model demonstrated robustness against data
poisoning attacks by selectively filtering out erroneous memories, leveraging
the Hebb repetition effect to reinforce the retention of significant data.</abstract><doi>10.48550/arxiv.2407.07133</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2407.07133 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2407_07133 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning Computer Science - Neural and Evolutionary Computing |
title | Neuromimetic metaplasticity for adaptive continual learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T12%3A20%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Neuromimetic%20metaplasticity%20for%20adaptive%20continual%20learning&rft.au=Cho,%20Suhee&rft.date=2024-07-09&rft_id=info:doi/10.48550/arxiv.2407.07133&rft_dat=%3Carxiv_GOX%3E2407_07133%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |