Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization

Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Proceedings of the National Academy of Sciences - PNAS 2018-10, Vol.115 (44), p.E10467-E10475
Hauptverfasser: Masse, Nicolas Y., Grant, Gregory D., Freedman, David J.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page E10475
container_issue 44
container_start_page E10467
container_title Proceedings of the National Academy of Sciences - PNAS
container_volume 115
creator Masse, Nicolas Y.
Grant, Gregory D.
Freedman, David J.
description Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.
doi_str_mv 10.1073/pnas.1803839115
format Article
fullrecord <record><control><sourceid>jstor_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_6217392</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><jstor_id>26532472</jstor_id><sourcerecordid>26532472</sourcerecordid><originalsourceid>FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</originalsourceid><addsrcrecordid>eNpdkc1rVDEUxYNY7Fhdu1IG3HTz2tx8ZyOU0qpQcKNLCZm8vGmGN8kzySutf33TTh2tqwv3_M7hXg5C7wCfAJb0dIq2nIDCVFENwF-gBWANnWAav0QLjInsFCPsEL0uZYMx1lzhV-iQYgocmFygn2fj6G-CrSGul85WW2pO03VwyyHlta-P-7k8qilWf1u73k8-9j7W5Xpns7Fflrtop9pspdpVGMPvJqX4Bh0Mdiz-7dM8Qj8uL76ff-muvn3-en521TlGde0UgIMBuBNSKOBWYswGzskge64EkcCI7fnKOst8U5x0xIGkQrueYSwHeoQ-7XKnebX1vWvHZTuaKYetzXcm2WCeKzFcm3W6MYK0HE1awPFTQE6_Zl-q2Ybi_Dja6NNcDAHQmgihRUM__odu0pxje69RFEsJStFGne4ol1Mp2Q_7YwCbh-rMQ3Xmb3XN8eHfH_b8n64a8H4HbEpNea8TwSlhktB7H8Sgsg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2130771883</pqid></control><display><type>article</type><title>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</title><source>Jstor Complete Legacy</source><source>MEDLINE</source><source>PubMed Central</source><source>Alma/SFX Local Collection</source><source>Free Full-Text Journals in Chemistry</source><creator>Masse, Nicolas Y. ; Grant, Gregory D. ; Freedman, David J.</creator><creatorcontrib>Masse, Nicolas Y. ; Grant, Gregory D. ; Freedman, David J.</creatorcontrib><description>Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.</description><identifier>ISSN: 0027-8424</identifier><identifier>EISSN: 1091-6490</identifier><identifier>DOI: 10.1073/pnas.1803839115</identifier><identifier>PMID: 30315147</identifier><language>eng</language><publisher>United States: National Academy of Sciences</publisher><subject>Algorithms ; Artificial intelligence ; Artificial neural networks ; Biological Sciences ; Brain ; Computational neuroscience ; Gating ; In vivo methods and tests ; Learning ; Learning theory ; Machine Learning ; Memory ; Neural networks ; Neural Networks (Computer) ; Performance degradation ; Physical Sciences ; PNAS Plus ; Stabilization ; Task Performance and Analysis ; Training ; Weight</subject><ispartof>Proceedings of the National Academy of Sciences - PNAS, 2018-10, Vol.115 (44), p.E10467-E10475</ispartof><rights>Volumes 1–89 and 106–115, copyright as a collective work only; author(s) retains copyright to individual articles</rights><rights>Copyright National Academy of Sciences Oct 30, 2018</rights><rights>2018</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</citedby><cites>FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</cites><orcidid>0000-0002-9094-1298</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.jstor.org/stable/pdf/26532472$$EPDF$$P50$$Gjstor$$H</linktopdf><linktohtml>$$Uhttps://www.jstor.org/stable/26532472$$EHTML$$P50$$Gjstor$$H</linktohtml><link.rule.ids>230,314,723,776,780,799,881,27901,27902,53766,53768,57992,58225</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/30315147$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Masse, Nicolas Y.</creatorcontrib><creatorcontrib>Grant, Gregory D.</creatorcontrib><creatorcontrib>Freedman, David J.</creatorcontrib><title>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</title><title>Proceedings of the National Academy of Sciences - PNAS</title><addtitle>Proc Natl Acad Sci U S A</addtitle><description>Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.</description><subject>Algorithms</subject><subject>Artificial intelligence</subject><subject>Artificial neural networks</subject><subject>Biological Sciences</subject><subject>Brain</subject><subject>Computational neuroscience</subject><subject>Gating</subject><subject>In vivo methods and tests</subject><subject>Learning</subject><subject>Learning theory</subject><subject>Machine Learning</subject><subject>Memory</subject><subject>Neural networks</subject><subject>Neural Networks (Computer)</subject><subject>Performance degradation</subject><subject>Physical Sciences</subject><subject>PNAS Plus</subject><subject>Stabilization</subject><subject>Task Performance and Analysis</subject><subject>Training</subject><subject>Weight</subject><issn>0027-8424</issn><issn>1091-6490</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNpdkc1rVDEUxYNY7Fhdu1IG3HTz2tx8ZyOU0qpQcKNLCZm8vGmGN8kzySutf33TTh2tqwv3_M7hXg5C7wCfAJb0dIq2nIDCVFENwF-gBWANnWAav0QLjInsFCPsEL0uZYMx1lzhV-iQYgocmFygn2fj6G-CrSGul85WW2pO03VwyyHlta-P-7k8qilWf1u73k8-9j7W5Xpns7Fflrtop9pspdpVGMPvJqX4Bh0Mdiz-7dM8Qj8uL76ff-muvn3-en521TlGde0UgIMBuBNSKOBWYswGzskge64EkcCI7fnKOst8U5x0xIGkQrueYSwHeoQ-7XKnebX1vWvHZTuaKYetzXcm2WCeKzFcm3W6MYK0HE1awPFTQE6_Zl-q2Ybi_Dja6NNcDAHQmgihRUM__odu0pxje69RFEsJStFGne4ol1Mp2Q_7YwCbh-rMQ3Xmb3XN8eHfH_b8n64a8H4HbEpNea8TwSlhktB7H8Sgsg</recordid><startdate>20181030</startdate><enddate>20181030</enddate><creator>Masse, Nicolas Y.</creator><creator>Grant, Gregory D.</creator><creator>Freedman, David J.</creator><general>National Academy of Sciences</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QG</scope><scope>7QL</scope><scope>7QP</scope><scope>7QR</scope><scope>7SN</scope><scope>7SS</scope><scope>7T5</scope><scope>7TK</scope><scope>7TM</scope><scope>7TO</scope><scope>7U9</scope><scope>8FD</scope><scope>C1K</scope><scope>FR3</scope><scope>H94</scope><scope>M7N</scope><scope>P64</scope><scope>RC3</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0002-9094-1298</orcidid></search><sort><creationdate>20181030</creationdate><title>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</title><author>Masse, Nicolas Y. ; Grant, Gregory D. ; Freedman, David J.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Algorithms</topic><topic>Artificial intelligence</topic><topic>Artificial neural networks</topic><topic>Biological Sciences</topic><topic>Brain</topic><topic>Computational neuroscience</topic><topic>Gating</topic><topic>In vivo methods and tests</topic><topic>Learning</topic><topic>Learning theory</topic><topic>Machine Learning</topic><topic>Memory</topic><topic>Neural networks</topic><topic>Neural Networks (Computer)</topic><topic>Performance degradation</topic><topic>Physical Sciences</topic><topic>PNAS Plus</topic><topic>Stabilization</topic><topic>Task Performance and Analysis</topic><topic>Training</topic><topic>Weight</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Masse, Nicolas Y.</creatorcontrib><creatorcontrib>Grant, Gregory D.</creatorcontrib><creatorcontrib>Freedman, David J.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Animal Behavior Abstracts</collection><collection>Bacteriology Abstracts (Microbiology B)</collection><collection>Calcium &amp; Calcified Tissue Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Immunology Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Nucleic Acids Abstracts</collection><collection>Oncogenes and Growth Factors Abstracts</collection><collection>Virology and AIDS Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>Engineering Research Database</collection><collection>AIDS and Cancer Research Abstracts</collection><collection>Algology Mycology and Protozoology Abstracts (Microbiology C)</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>Genetics Abstracts</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Masse, Nicolas Y.</au><au>Grant, Gregory D.</au><au>Freedman, David J.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</atitle><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle><addtitle>Proc Natl Acad Sci U S A</addtitle><date>2018-10-30</date><risdate>2018</risdate><volume>115</volume><issue>44</issue><spage>E10467</spage><epage>E10475</epage><pages>E10467-E10475</pages><issn>0027-8424</issn><eissn>1091-6490</eissn><abstract>Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.</abstract><cop>United States</cop><pub>National Academy of Sciences</pub><pmid>30315147</pmid><doi>10.1073/pnas.1803839115</doi><orcidid>https://orcid.org/0000-0002-9094-1298</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0027-8424
ispartof Proceedings of the National Academy of Sciences - PNAS, 2018-10, Vol.115 (44), p.E10467-E10475
issn 0027-8424
1091-6490
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_6217392
source Jstor Complete Legacy; MEDLINE; PubMed Central; Alma/SFX Local Collection; Free Full-Text Journals in Chemistry
subjects Algorithms
Artificial intelligence
Artificial neural networks
Biological Sciences
Brain
Computational neuroscience
Gating
In vivo methods and tests
Learning
Learning theory
Machine Learning
Memory
Neural networks
Neural Networks (Computer)
Performance degradation
Physical Sciences
PNAS Plus
Stabilization
Task Performance and Analysis
Training
Weight
title Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-03T16%3A58%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-jstor_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Alleviating%20catastrophic%20forgetting%20using%20context-dependent%20gating%20and%20synaptic%20stabilization&rft.jtitle=Proceedings%20of%20the%20National%20Academy%20of%20Sciences%20-%20PNAS&rft.au=Masse,%20Nicolas%20Y.&rft.date=2018-10-30&rft.volume=115&rft.issue=44&rft.spage=E10467&rft.epage=E10475&rft.pages=E10467-E10475&rft.issn=0027-8424&rft.eissn=1091-6490&rft_id=info:doi/10.1073/pnas.1803839115&rft_dat=%3Cjstor_pubme%3E26532472%3C/jstor_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2130771883&rft_id=info:pmid/30315147&rft_jstor_id=26532472&rfr_iscdi=true