Discovering and Explaining the Representation Bottleneck of DNNs

This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the c...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Deng, Huiqi, Ren, Qihan, Zhang, Hao, Zhang, Quanshi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Deng, Huiqi
Ren, Qihan
Zhang, Hao
Zhang, Quanshi
description This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the complexity of interactions. We discover that a DNN is more likely to encode both too simple interactions and too complex interactions, but usually fails to learn interactions of intermediate complexity. Such a phenomenon is widely shared by different DNNs for different tasks. This phenomenon indicates a cognition gap between DNNs and human beings, and we call it a representation bottleneck. We theoretically prove the underlying reason for the representation bottleneck. Furthermore, we propose a loss to encourage/penalize the learning of interactions of specific complexities, and analyze the representation capacities of interactions of different complexities.
doi_str_mv 10.48550/arxiv.2111.06236
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2111_06236</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2111_06236</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-687a6f43492ee47e08b1539e8ab136e0c8eb181369f7b60eaddd586c8242b9f33</originalsourceid><addsrcrecordid>eNotz8tOwzAUBFBvWKDCB7DCP5DgV26cHdCWh1QVCXUfXSfXYBGcyLGq8vfQ0tXMbEY6jN1IURpbVeIO0yHsSyWlLAUoDZfsfhXmbtxTCvGDY-z5-jANGOJx5k_i7zQlmilmzGGM_HHMeaBI3RcfPV9tt_MVu_A4zHR9zgXbPa13y5di8_b8unzYFAg1FGBrBG-0aRSRqUlYJyvdkEUnNZDoLDlp_2rjaweCsO_7ykJnlVGu8Vov2O3_7YnQTil8Y_ppj5T2RNG__-hDhQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Discovering and Explaining the Representation Bottleneck of DNNs</title><source>arXiv.org</source><creator>Deng, Huiqi ; Ren, Qihan ; Zhang, Hao ; Zhang, Quanshi</creator><creatorcontrib>Deng, Huiqi ; Ren, Qihan ; Zhang, Hao ; Zhang, Quanshi</creatorcontrib><description>This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the complexity of interactions. We discover that a DNN is more likely to encode both too simple interactions and too complex interactions, but usually fails to learn interactions of intermediate complexity. Such a phenomenon is widely shared by different DNNs for different tasks. This phenomenon indicates a cognition gap between DNNs and human beings, and we call it a representation bottleneck. We theoretically prove the underlying reason for the representation bottleneck. Furthermore, we propose a loss to encourage/penalize the learning of interactions of specific complexities, and analyze the representation capacities of interactions of different complexities.</description><identifier>DOI: 10.48550/arxiv.2111.06236</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2021-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2111.06236$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2111.06236$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Deng, Huiqi</creatorcontrib><creatorcontrib>Ren, Qihan</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Zhang, Quanshi</creatorcontrib><title>Discovering and Explaining the Representation Bottleneck of DNNs</title><description>This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the complexity of interactions. We discover that a DNN is more likely to encode both too simple interactions and too complex interactions, but usually fails to learn interactions of intermediate complexity. Such a phenomenon is widely shared by different DNNs for different tasks. This phenomenon indicates a cognition gap between DNNs and human beings, and we call it a representation bottleneck. We theoretically prove the underlying reason for the representation bottleneck. Furthermore, we propose a loss to encourage/penalize the learning of interactions of specific complexities, and analyze the representation capacities of interactions of different complexities.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz8tOwzAUBFBvWKDCB7DCP5DgV26cHdCWh1QVCXUfXSfXYBGcyLGq8vfQ0tXMbEY6jN1IURpbVeIO0yHsSyWlLAUoDZfsfhXmbtxTCvGDY-z5-jANGOJx5k_i7zQlmilmzGGM_HHMeaBI3RcfPV9tt_MVu_A4zHR9zgXbPa13y5di8_b8unzYFAg1FGBrBG-0aRSRqUlYJyvdkEUnNZDoLDlp_2rjaweCsO_7ykJnlVGu8Vov2O3_7YnQTil8Y_ppj5T2RNG__-hDhQ</recordid><startdate>20211111</startdate><enddate>20211111</enddate><creator>Deng, Huiqi</creator><creator>Ren, Qihan</creator><creator>Zhang, Hao</creator><creator>Zhang, Quanshi</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20211111</creationdate><title>Discovering and Explaining the Representation Bottleneck of DNNs</title><author>Deng, Huiqi ; Ren, Qihan ; Zhang, Hao ; Zhang, Quanshi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-687a6f43492ee47e08b1539e8ab136e0c8eb181369f7b60eaddd586c8242b9f33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Deng, Huiqi</creatorcontrib><creatorcontrib>Ren, Qihan</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Zhang, Quanshi</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Deng, Huiqi</au><au>Ren, Qihan</au><au>Zhang, Hao</au><au>Zhang, Quanshi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Discovering and Explaining the Representation Bottleneck of DNNs</atitle><date>2021-11-11</date><risdate>2021</risdate><abstract>This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the complexity of interactions. We discover that a DNN is more likely to encode both too simple interactions and too complex interactions, but usually fails to learn interactions of intermediate complexity. Such a phenomenon is widely shared by different DNNs for different tasks. This phenomenon indicates a cognition gap between DNNs and human beings, and we call it a representation bottleneck. We theoretically prove the underlying reason for the representation bottleneck. Furthermore, we propose a loss to encourage/penalize the learning of interactions of specific complexities, and analyze the representation capacities of interactions of different complexities.</abstract><doi>10.48550/arxiv.2111.06236</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2111.06236
ispartof
issn
language eng
recordid cdi_arxiv_primary_2111_06236
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
title Discovering and Explaining the Representation Bottleneck of DNNs
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T05%3A02%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Discovering%20and%20Explaining%20the%20Representation%20Bottleneck%20of%20DNNs&rft.au=Deng,%20Huiqi&rft.date=2021-11-11&rft_id=info:doi/10.48550/arxiv.2111.06236&rft_dat=%3Carxiv_GOX%3E2111_06236%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true