Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization
Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection...
Gespeichert in:
Veröffentlicht in: | Proceedings of the National Academy of Sciences - PNAS 2018-10, Vol.115 (44), p.E10467-E10475 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | E10475 |
---|---|
container_issue | 44 |
container_start_page | E10467 |
container_title | Proceedings of the National Academy of Sciences - PNAS |
container_volume | 115 |
creator | Masse, Nicolas Y. Grant, Gregory D. Freedman, David J. |
description | Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning. |
doi_str_mv | 10.1073/pnas.1803839115 |
format | Article |
fullrecord | <record><control><sourceid>jstor_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_6217392</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><jstor_id>26532472</jstor_id><sourcerecordid>26532472</sourcerecordid><originalsourceid>FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</originalsourceid><addsrcrecordid>eNpdkc1rVDEUxYNY7Fhdu1IG3HTz2tx8ZyOU0qpQcKNLCZm8vGmGN8kzySutf33TTh2tqwv3_M7hXg5C7wCfAJb0dIq2nIDCVFENwF-gBWANnWAav0QLjInsFCPsEL0uZYMx1lzhV-iQYgocmFygn2fj6G-CrSGul85WW2pO03VwyyHlta-P-7k8qilWf1u73k8-9j7W5Xpns7Fflrtop9pspdpVGMPvJqX4Bh0Mdiz-7dM8Qj8uL76ff-muvn3-en521TlGde0UgIMBuBNSKOBWYswGzskge64EkcCI7fnKOst8U5x0xIGkQrueYSwHeoQ-7XKnebX1vWvHZTuaKYetzXcm2WCeKzFcm3W6MYK0HE1awPFTQE6_Zl-q2Ybi_Dja6NNcDAHQmgihRUM__odu0pxje69RFEsJStFGne4ol1Mp2Q_7YwCbh-rMQ3Xmb3XN8eHfH_b8n64a8H4HbEpNea8TwSlhktB7H8Sgsg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2130771883</pqid></control><display><type>article</type><title>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</title><source>Jstor Complete Legacy</source><source>MEDLINE</source><source>PubMed Central</source><source>Alma/SFX Local Collection</source><source>Free Full-Text Journals in Chemistry</source><creator>Masse, Nicolas Y. ; Grant, Gregory D. ; Freedman, David J.</creator><creatorcontrib>Masse, Nicolas Y. ; Grant, Gregory D. ; Freedman, David J.</creatorcontrib><description>Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.</description><identifier>ISSN: 0027-8424</identifier><identifier>EISSN: 1091-6490</identifier><identifier>DOI: 10.1073/pnas.1803839115</identifier><identifier>PMID: 30315147</identifier><language>eng</language><publisher>United States: National Academy of Sciences</publisher><subject>Algorithms ; Artificial intelligence ; Artificial neural networks ; Biological Sciences ; Brain ; Computational neuroscience ; Gating ; In vivo methods and tests ; Learning ; Learning theory ; Machine Learning ; Memory ; Neural networks ; Neural Networks (Computer) ; Performance degradation ; Physical Sciences ; PNAS Plus ; Stabilization ; Task Performance and Analysis ; Training ; Weight</subject><ispartof>Proceedings of the National Academy of Sciences - PNAS, 2018-10, Vol.115 (44), p.E10467-E10475</ispartof><rights>Volumes 1–89 and 106–115, copyright as a collective work only; author(s) retains copyright to individual articles</rights><rights>Copyright National Academy of Sciences Oct 30, 2018</rights><rights>2018</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</citedby><cites>FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</cites><orcidid>0000-0002-9094-1298</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.jstor.org/stable/pdf/26532472$$EPDF$$P50$$Gjstor$$H</linktopdf><linktohtml>$$Uhttps://www.jstor.org/stable/26532472$$EHTML$$P50$$Gjstor$$H</linktohtml><link.rule.ids>230,314,723,776,780,799,881,27901,27902,53766,53768,57992,58225</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/30315147$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Masse, Nicolas Y.</creatorcontrib><creatorcontrib>Grant, Gregory D.</creatorcontrib><creatorcontrib>Freedman, David J.</creatorcontrib><title>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</title><title>Proceedings of the National Academy of Sciences - PNAS</title><addtitle>Proc Natl Acad Sci U S A</addtitle><description>Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.</description><subject>Algorithms</subject><subject>Artificial intelligence</subject><subject>Artificial neural networks</subject><subject>Biological Sciences</subject><subject>Brain</subject><subject>Computational neuroscience</subject><subject>Gating</subject><subject>In vivo methods and tests</subject><subject>Learning</subject><subject>Learning theory</subject><subject>Machine Learning</subject><subject>Memory</subject><subject>Neural networks</subject><subject>Neural Networks (Computer)</subject><subject>Performance degradation</subject><subject>Physical Sciences</subject><subject>PNAS Plus</subject><subject>Stabilization</subject><subject>Task Performance and Analysis</subject><subject>Training</subject><subject>Weight</subject><issn>0027-8424</issn><issn>1091-6490</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNpdkc1rVDEUxYNY7Fhdu1IG3HTz2tx8ZyOU0qpQcKNLCZm8vGmGN8kzySutf33TTh2tqwv3_M7hXg5C7wCfAJb0dIq2nIDCVFENwF-gBWANnWAav0QLjInsFCPsEL0uZYMx1lzhV-iQYgocmFygn2fj6G-CrSGul85WW2pO03VwyyHlta-P-7k8qilWf1u73k8-9j7W5Xpns7Fflrtop9pspdpVGMPvJqX4Bh0Mdiz-7dM8Qj8uL76ff-muvn3-en521TlGde0UgIMBuBNSKOBWYswGzskge64EkcCI7fnKOst8U5x0xIGkQrueYSwHeoQ-7XKnebX1vWvHZTuaKYetzXcm2WCeKzFcm3W6MYK0HE1awPFTQE6_Zl-q2Ybi_Dja6NNcDAHQmgihRUM__odu0pxje69RFEsJStFGne4ol1Mp2Q_7YwCbh-rMQ3Xmb3XN8eHfH_b8n64a8H4HbEpNea8TwSlhktB7H8Sgsg</recordid><startdate>20181030</startdate><enddate>20181030</enddate><creator>Masse, Nicolas Y.</creator><creator>Grant, Gregory D.</creator><creator>Freedman, David J.</creator><general>National Academy of Sciences</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QG</scope><scope>7QL</scope><scope>7QP</scope><scope>7QR</scope><scope>7SN</scope><scope>7SS</scope><scope>7T5</scope><scope>7TK</scope><scope>7TM</scope><scope>7TO</scope><scope>7U9</scope><scope>8FD</scope><scope>C1K</scope><scope>FR3</scope><scope>H94</scope><scope>M7N</scope><scope>P64</scope><scope>RC3</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0002-9094-1298</orcidid></search><sort><creationdate>20181030</creationdate><title>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</title><author>Masse, Nicolas Y. ; Grant, Gregory D. ; Freedman, David J.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c439t-811c1f15c676815a7004f552f7d58627142ad5baca4e04fc7c2c17369cd4007f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Algorithms</topic><topic>Artificial intelligence</topic><topic>Artificial neural networks</topic><topic>Biological Sciences</topic><topic>Brain</topic><topic>Computational neuroscience</topic><topic>Gating</topic><topic>In vivo methods and tests</topic><topic>Learning</topic><topic>Learning theory</topic><topic>Machine Learning</topic><topic>Memory</topic><topic>Neural networks</topic><topic>Neural Networks (Computer)</topic><topic>Performance degradation</topic><topic>Physical Sciences</topic><topic>PNAS Plus</topic><topic>Stabilization</topic><topic>Task Performance and Analysis</topic><topic>Training</topic><topic>Weight</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Masse, Nicolas Y.</creatorcontrib><creatorcontrib>Grant, Gregory D.</creatorcontrib><creatorcontrib>Freedman, David J.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Animal Behavior Abstracts</collection><collection>Bacteriology Abstracts (Microbiology B)</collection><collection>Calcium & Calcified Tissue Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Immunology Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Nucleic Acids Abstracts</collection><collection>Oncogenes and Growth Factors Abstracts</collection><collection>Virology and AIDS Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>Engineering Research Database</collection><collection>AIDS and Cancer Research Abstracts</collection><collection>Algology Mycology and Protozoology Abstracts (Microbiology C)</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>Genetics Abstracts</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Masse, Nicolas Y.</au><au>Grant, Gregory D.</au><au>Freedman, David J.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization</atitle><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle><addtitle>Proc Natl Acad Sci U S A</addtitle><date>2018-10-30</date><risdate>2018</risdate><volume>115</volume><issue>44</issue><spage>E10467</spage><epage>E10475</epage><pages>E10467-E10475</pages><issn>0027-8424</issn><eissn>1091-6490</eissn><abstract>Humans and most animals can learn new tasks without forgetting old ones. However, training artificial neural networks (ANNs) on new tasks typically causes them to forget previously learned tasks. This phenomenon is the result of “catastrophic forgetting,” in which training an ANN disrupts connection weights that were important for solving previous tasks, degrading task performance. Several recent studies have proposed methods to stabilize connection weights of ANNs that are deemed most important for solving a task, which helps alleviate catastrophic forgetting. Here, drawing inspiration from algorithms that are believed to be implemented in vivo, we propose a complementary method: adding a context-dependent gating signal, such that only sparse, mostly nonoverlapping patterns of units are active for any one task. This method is easy to implement, requires little computational overhead, and allows ANNs to maintain high performance across large numbers of sequentially presented tasks, particularly when combined with weight stabilization. We show that this method works for both feedforward and recurrent network architectures, trained using either supervised or reinforcement-based learning. This suggests that using multiple, complementary methods, akin to what is believed to occur in the brain, can be a highly effective strategy to support continual learning.</abstract><cop>United States</cop><pub>National Academy of Sciences</pub><pmid>30315147</pmid><doi>10.1073/pnas.1803839115</doi><orcidid>https://orcid.org/0000-0002-9094-1298</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0027-8424 |
ispartof | Proceedings of the National Academy of Sciences - PNAS, 2018-10, Vol.115 (44), p.E10467-E10475 |
issn | 0027-8424 1091-6490 |
language | eng |
recordid | cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_6217392 |
source | Jstor Complete Legacy; MEDLINE; PubMed Central; Alma/SFX Local Collection; Free Full-Text Journals in Chemistry |
subjects | Algorithms Artificial intelligence Artificial neural networks Biological Sciences Brain Computational neuroscience Gating In vivo methods and tests Learning Learning theory Machine Learning Memory Neural networks Neural Networks (Computer) Performance degradation Physical Sciences PNAS Plus Stabilization Task Performance and Analysis Training Weight |
title | Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-03T16%3A58%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-jstor_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Alleviating%20catastrophic%20forgetting%20using%20context-dependent%20gating%20and%20synaptic%20stabilization&rft.jtitle=Proceedings%20of%20the%20National%20Academy%20of%20Sciences%20-%20PNAS&rft.au=Masse,%20Nicolas%20Y.&rft.date=2018-10-30&rft.volume=115&rft.issue=44&rft.spage=E10467&rft.epage=E10475&rft.pages=E10467-E10475&rft.issn=0027-8424&rft.eissn=1091-6490&rft_id=info:doi/10.1073/pnas.1803839115&rft_dat=%3Cjstor_pubme%3E26532472%3C/jstor_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2130771883&rft_id=info:pmid/30315147&rft_jstor_id=26532472&rfr_iscdi=true |