Three approaches to facilitate invariant neurons and generalization to out-of-distribution orientations and illuminations

The training data distribution is often biased towards objects in certain orientations and illumination conditions. While humans have a remarkable capability of recognizing objects in out-of-distribution (OoD) orientations and illuminations, Deep Neural Networks (DNNs) severely suffer in this case,...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Neural networks 2022-11, Vol.155, p.119-143
Hauptverfasser: Sakai, Akira, Sunagawa, Taro, Madan, Spandan, Suzuki, Kanata, Katoh, Takashi, Kobashi, Hiromichi, Pfister, Hanspeter, Sinha, Pawan, Boix, Xavier, Sasaki, Tomotake
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 143
container_issue
container_start_page 119
container_title Neural networks
container_volume 155
creator Sakai, Akira
Sunagawa, Taro
Madan, Spandan
Suzuki, Kanata
Katoh, Takashi
Kobashi, Hiromichi
Pfister, Hanspeter
Sinha, Pawan
Boix, Xavier
Sasaki, Tomotake
description The training data distribution is often biased towards objects in certain orientations and illumination conditions. While humans have a remarkable capability of recognizing objects in out-of-distribution (OoD) orientations and illuminations, Deep Neural Networks (DNNs) severely suffer in this case, even when large amounts of training examples are available. Neurons that are invariant to orientations and illuminations have been proposed as a neural mechanism that could facilitate OoD generalization, but it is unclear how to encourage the emergence of such invariant neurons. In this paper, we investigate three different approaches that lead to the emergence of invariant neurons and substantially improve DNNs in recognizing objects in OoD orientations and illuminations. Namely, these approaches are (i) training much longer after convergence of the in-distribution (InD) validation accuracy, i.e., late-stopping, (ii) tuning the momentum parameter of the batch normalization layers, and (iii) enforcing invariance of the neural activity in an intermediate layer to orientation and illumination conditions. Each of these approaches substantially improves the DNN’s OoD accuracy (more than 20% in some cases). We report results in four datasets: two datasets are modified from the MNIST and iLab datasets, and the other two are novel (one of 3D rendered cars and another of objects taken from various controlled orientations and illumination conditions). These datasets allow to study the effects of different amounts of bias and are challenging as DNNs perform poorly in OoD conditions. Finally, we demonstrate that even though the three approaches focus on different aspects of DNNs, they all tend to lead to the same underlying neural mechanism to enable OoD accuracy gains — individual neurons in the intermediate layers become invariant to OoD orientations and illuminations. We anticipate this study to be a basis for further improvement of deep neural networks’ OoD generalization performance, which is highly demanded to achieve safe and fair AI applications. •We investigate three approaches to enhance out-of-distribution (OoD) generalization.•We provide four datasets for the research of OoD generalization.•We study the individual neuron level of selectivity and invariance experimentally.•This mechanism drives the improvements of the OoD accuracy of the three approaches.
doi_str_mv 10.1016/j.neunet.2022.07.026
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2709739966</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S089360802200288X</els_id><sourcerecordid>2709739966</sourcerecordid><originalsourceid>FETCH-LOGICAL-c385t-b79482a5706c31725d4dc7b0e2946abe5b974e9bd19a50743479894a0534ac3a3</originalsourceid><addsrcrecordid>eNp9kEGL2zAQhcWyhc1u-w968LEXu2NJtqzLQgntdiGwl_QsxvKkUXCkrCQH0l9fJ-65p4HH-97wHmOfa6hqqNuvh8rT5ClXHDivQFXA2zu2qjulS646fs9W0GlRttDBA3tM6QAAbSfFil22-0hU4OkUA9o9pSKHYofWjS5jpsL5M0aHPhfzixh8KtAPxW_yFHF0fzC74K9ImHIZduXgUo6un25yiI58vlkWzI3jdHR-UT6yDzscE336d5_Yrx_ft-uf5ebt5XX9bVNa0TW57JWWHcdGQWtFrXgzyMGqHohr2WJPTa-VJN0PtcYGlBRS6U5LhEZItALFE_uy5M4N3ydK2RxdsjSO6ClMyXAFWgmt23a2ysVqY0gp0s6cojtivJgazHVpczDL0ua6tAFl5qVn7HnBaK5xdhRNsnNzS4OLZLMZgvt_wF_cdoxe</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2709739966</pqid></control><display><type>article</type><title>Three approaches to facilitate invariant neurons and generalization to out-of-distribution orientations and illuminations</title><source>Elsevier ScienceDirect Journals</source><creator>Sakai, Akira ; Sunagawa, Taro ; Madan, Spandan ; Suzuki, Kanata ; Katoh, Takashi ; Kobashi, Hiromichi ; Pfister, Hanspeter ; Sinha, Pawan ; Boix, Xavier ; Sasaki, Tomotake</creator><creatorcontrib>Sakai, Akira ; Sunagawa, Taro ; Madan, Spandan ; Suzuki, Kanata ; Katoh, Takashi ; Kobashi, Hiromichi ; Pfister, Hanspeter ; Sinha, Pawan ; Boix, Xavier ; Sasaki, Tomotake</creatorcontrib><description>The training data distribution is often biased towards objects in certain orientations and illumination conditions. While humans have a remarkable capability of recognizing objects in out-of-distribution (OoD) orientations and illuminations, Deep Neural Networks (DNNs) severely suffer in this case, even when large amounts of training examples are available. Neurons that are invariant to orientations and illuminations have been proposed as a neural mechanism that could facilitate OoD generalization, but it is unclear how to encourage the emergence of such invariant neurons. In this paper, we investigate three different approaches that lead to the emergence of invariant neurons and substantially improve DNNs in recognizing objects in OoD orientations and illuminations. Namely, these approaches are (i) training much longer after convergence of the in-distribution (InD) validation accuracy, i.e., late-stopping, (ii) tuning the momentum parameter of the batch normalization layers, and (iii) enforcing invariance of the neural activity in an intermediate layer to orientation and illumination conditions. Each of these approaches substantially improves the DNN’s OoD accuracy (more than 20% in some cases). We report results in four datasets: two datasets are modified from the MNIST and iLab datasets, and the other two are novel (one of 3D rendered cars and another of objects taken from various controlled orientations and illumination conditions). These datasets allow to study the effects of different amounts of bias and are challenging as DNNs perform poorly in OoD conditions. Finally, we demonstrate that even though the three approaches focus on different aspects of DNNs, they all tend to lead to the same underlying neural mechanism to enable OoD accuracy gains — individual neurons in the intermediate layers become invariant to OoD orientations and illuminations. We anticipate this study to be a basis for further improvement of deep neural networks’ OoD generalization performance, which is highly demanded to achieve safe and fair AI applications. •We investigate three approaches to enhance out-of-distribution (OoD) generalization.•We provide four datasets for the research of OoD generalization.•We study the individual neuron level of selectivity and invariance experimentally.•This mechanism drives the improvements of the OoD accuracy of the three approaches.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/j.neunet.2022.07.026</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Neural activity analysis ; Neural invariance ; Neural selectivity ; Object recognition in novel conditions ; Out-of-distribution generalization</subject><ispartof>Neural networks, 2022-11, Vol.155, p.119-143</ispartof><rights>2022 The Authors</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c385t-b79482a5706c31725d4dc7b0e2946abe5b974e9bd19a50743479894a0534ac3a3</citedby><cites>FETCH-LOGICAL-c385t-b79482a5706c31725d4dc7b0e2946abe5b974e9bd19a50743479894a0534ac3a3</cites><orcidid>0000-0002-3376-2779 ; 0000-0002-5543-9542 ; 0000-0003-4656-3485 ; 0000-0001-7122-7649 ; 0000-0002-8957-0004 ; 0000-0002-3620-2582 ; 0000-0003-3951-1089</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S089360802200288X$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids></links><search><creatorcontrib>Sakai, Akira</creatorcontrib><creatorcontrib>Sunagawa, Taro</creatorcontrib><creatorcontrib>Madan, Spandan</creatorcontrib><creatorcontrib>Suzuki, Kanata</creatorcontrib><creatorcontrib>Katoh, Takashi</creatorcontrib><creatorcontrib>Kobashi, Hiromichi</creatorcontrib><creatorcontrib>Pfister, Hanspeter</creatorcontrib><creatorcontrib>Sinha, Pawan</creatorcontrib><creatorcontrib>Boix, Xavier</creatorcontrib><creatorcontrib>Sasaki, Tomotake</creatorcontrib><title>Three approaches to facilitate invariant neurons and generalization to out-of-distribution orientations and illuminations</title><title>Neural networks</title><description>The training data distribution is often biased towards objects in certain orientations and illumination conditions. While humans have a remarkable capability of recognizing objects in out-of-distribution (OoD) orientations and illuminations, Deep Neural Networks (DNNs) severely suffer in this case, even when large amounts of training examples are available. Neurons that are invariant to orientations and illuminations have been proposed as a neural mechanism that could facilitate OoD generalization, but it is unclear how to encourage the emergence of such invariant neurons. In this paper, we investigate three different approaches that lead to the emergence of invariant neurons and substantially improve DNNs in recognizing objects in OoD orientations and illuminations. Namely, these approaches are (i) training much longer after convergence of the in-distribution (InD) validation accuracy, i.e., late-stopping, (ii) tuning the momentum parameter of the batch normalization layers, and (iii) enforcing invariance of the neural activity in an intermediate layer to orientation and illumination conditions. Each of these approaches substantially improves the DNN’s OoD accuracy (more than 20% in some cases). We report results in four datasets: two datasets are modified from the MNIST and iLab datasets, and the other two are novel (one of 3D rendered cars and another of objects taken from various controlled orientations and illumination conditions). These datasets allow to study the effects of different amounts of bias and are challenging as DNNs perform poorly in OoD conditions. Finally, we demonstrate that even though the three approaches focus on different aspects of DNNs, they all tend to lead to the same underlying neural mechanism to enable OoD accuracy gains — individual neurons in the intermediate layers become invariant to OoD orientations and illuminations. We anticipate this study to be a basis for further improvement of deep neural networks’ OoD generalization performance, which is highly demanded to achieve safe and fair AI applications. •We investigate three approaches to enhance out-of-distribution (OoD) generalization.•We provide four datasets for the research of OoD generalization.•We study the individual neuron level of selectivity and invariance experimentally.•This mechanism drives the improvements of the OoD accuracy of the three approaches.</description><subject>Neural activity analysis</subject><subject>Neural invariance</subject><subject>Neural selectivity</subject><subject>Object recognition in novel conditions</subject><subject>Out-of-distribution generalization</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kEGL2zAQhcWyhc1u-w968LEXu2NJtqzLQgntdiGwl_QsxvKkUXCkrCQH0l9fJ-65p4HH-97wHmOfa6hqqNuvh8rT5ClXHDivQFXA2zu2qjulS646fs9W0GlRttDBA3tM6QAAbSfFil22-0hU4OkUA9o9pSKHYofWjS5jpsL5M0aHPhfzixh8KtAPxW_yFHF0fzC74K9ImHIZduXgUo6un25yiI58vlkWzI3jdHR-UT6yDzscE336d5_Yrx_ft-uf5ebt5XX9bVNa0TW57JWWHcdGQWtFrXgzyMGqHohr2WJPTa-VJN0PtcYGlBRS6U5LhEZItALFE_uy5M4N3ydK2RxdsjSO6ClMyXAFWgmt23a2ysVqY0gp0s6cojtivJgazHVpczDL0ua6tAFl5qVn7HnBaK5xdhRNsnNzS4OLZLMZgvt_wF_cdoxe</recordid><startdate>202211</startdate><enddate>202211</enddate><creator>Sakai, Akira</creator><creator>Sunagawa, Taro</creator><creator>Madan, Spandan</creator><creator>Suzuki, Kanata</creator><creator>Katoh, Takashi</creator><creator>Kobashi, Hiromichi</creator><creator>Pfister, Hanspeter</creator><creator>Sinha, Pawan</creator><creator>Boix, Xavier</creator><creator>Sasaki, Tomotake</creator><general>Elsevier Ltd</general><scope>6I.</scope><scope>AAFTH</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-3376-2779</orcidid><orcidid>https://orcid.org/0000-0002-5543-9542</orcidid><orcidid>https://orcid.org/0000-0003-4656-3485</orcidid><orcidid>https://orcid.org/0000-0001-7122-7649</orcidid><orcidid>https://orcid.org/0000-0002-8957-0004</orcidid><orcidid>https://orcid.org/0000-0002-3620-2582</orcidid><orcidid>https://orcid.org/0000-0003-3951-1089</orcidid></search><sort><creationdate>202211</creationdate><title>Three approaches to facilitate invariant neurons and generalization to out-of-distribution orientations and illuminations</title><author>Sakai, Akira ; Sunagawa, Taro ; Madan, Spandan ; Suzuki, Kanata ; Katoh, Takashi ; Kobashi, Hiromichi ; Pfister, Hanspeter ; Sinha, Pawan ; Boix, Xavier ; Sasaki, Tomotake</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c385t-b79482a5706c31725d4dc7b0e2946abe5b974e9bd19a50743479894a0534ac3a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Neural activity analysis</topic><topic>Neural invariance</topic><topic>Neural selectivity</topic><topic>Object recognition in novel conditions</topic><topic>Out-of-distribution generalization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Sakai, Akira</creatorcontrib><creatorcontrib>Sunagawa, Taro</creatorcontrib><creatorcontrib>Madan, Spandan</creatorcontrib><creatorcontrib>Suzuki, Kanata</creatorcontrib><creatorcontrib>Katoh, Takashi</creatorcontrib><creatorcontrib>Kobashi, Hiromichi</creatorcontrib><creatorcontrib>Pfister, Hanspeter</creatorcontrib><creatorcontrib>Sinha, Pawan</creatorcontrib><creatorcontrib>Boix, Xavier</creatorcontrib><creatorcontrib>Sasaki, Tomotake</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Sakai, Akira</au><au>Sunagawa, Taro</au><au>Madan, Spandan</au><au>Suzuki, Kanata</au><au>Katoh, Takashi</au><au>Kobashi, Hiromichi</au><au>Pfister, Hanspeter</au><au>Sinha, Pawan</au><au>Boix, Xavier</au><au>Sasaki, Tomotake</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Three approaches to facilitate invariant neurons and generalization to out-of-distribution orientations and illuminations</atitle><jtitle>Neural networks</jtitle><date>2022-11</date><risdate>2022</risdate><volume>155</volume><spage>119</spage><epage>143</epage><pages>119-143</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>The training data distribution is often biased towards objects in certain orientations and illumination conditions. While humans have a remarkable capability of recognizing objects in out-of-distribution (OoD) orientations and illuminations, Deep Neural Networks (DNNs) severely suffer in this case, even when large amounts of training examples are available. Neurons that are invariant to orientations and illuminations have been proposed as a neural mechanism that could facilitate OoD generalization, but it is unclear how to encourage the emergence of such invariant neurons. In this paper, we investigate three different approaches that lead to the emergence of invariant neurons and substantially improve DNNs in recognizing objects in OoD orientations and illuminations. Namely, these approaches are (i) training much longer after convergence of the in-distribution (InD) validation accuracy, i.e., late-stopping, (ii) tuning the momentum parameter of the batch normalization layers, and (iii) enforcing invariance of the neural activity in an intermediate layer to orientation and illumination conditions. Each of these approaches substantially improves the DNN’s OoD accuracy (more than 20% in some cases). We report results in four datasets: two datasets are modified from the MNIST and iLab datasets, and the other two are novel (one of 3D rendered cars and another of objects taken from various controlled orientations and illumination conditions). These datasets allow to study the effects of different amounts of bias and are challenging as DNNs perform poorly in OoD conditions. Finally, we demonstrate that even though the three approaches focus on different aspects of DNNs, they all tend to lead to the same underlying neural mechanism to enable OoD accuracy gains — individual neurons in the intermediate layers become invariant to OoD orientations and illuminations. We anticipate this study to be a basis for further improvement of deep neural networks’ OoD generalization performance, which is highly demanded to achieve safe and fair AI applications. •We investigate three approaches to enhance out-of-distribution (OoD) generalization.•We provide four datasets for the research of OoD generalization.•We study the individual neuron level of selectivity and invariance experimentally.•This mechanism drives the improvements of the OoD accuracy of the three approaches.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.neunet.2022.07.026</doi><tpages>25</tpages><orcidid>https://orcid.org/0000-0002-3376-2779</orcidid><orcidid>https://orcid.org/0000-0002-5543-9542</orcidid><orcidid>https://orcid.org/0000-0003-4656-3485</orcidid><orcidid>https://orcid.org/0000-0001-7122-7649</orcidid><orcidid>https://orcid.org/0000-0002-8957-0004</orcidid><orcidid>https://orcid.org/0000-0002-3620-2582</orcidid><orcidid>https://orcid.org/0000-0003-3951-1089</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0893-6080
ispartof Neural networks, 2022-11, Vol.155, p.119-143
issn 0893-6080
1879-2782
language eng
recordid cdi_proquest_miscellaneous_2709739966
source Elsevier ScienceDirect Journals
subjects Neural activity analysis
Neural invariance
Neural selectivity
Object recognition in novel conditions
Out-of-distribution generalization
title Three approaches to facilitate invariant neurons and generalization to out-of-distribution orientations and illuminations
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-03T02%3A19%3A15IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Three%20approaches%20to%20facilitate%20invariant%20neurons%20and%20generalization%20to%20out-of-distribution%20orientations%20and%20illuminations&rft.jtitle=Neural%20networks&rft.au=Sakai,%20Akira&rft.date=2022-11&rft.volume=155&rft.spage=119&rft.epage=143&rft.pages=119-143&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/j.neunet.2022.07.026&rft_dat=%3Cproquest_cross%3E2709739966%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2709739966&rft_id=info:pmid/&rft_els_id=S089360802200288X&rfr_iscdi=true