Discovering and Explaining the Representation Bottleneck of DNNs
This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the c...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Deng, Huiqi Ren, Qihan Zhang, Hao Zhang, Quanshi |
description | This paper explores the bottleneck of feature representations of deep neural
networks (DNNs), from the perspective of the complexity of interactions between
input variables encoded in DNNs. To this end, we focus on the multi-order
interaction between input variables, where the order represents the complexity
of interactions. We discover that a DNN is more likely to encode both too
simple interactions and too complex interactions, but usually fails to learn
interactions of intermediate complexity. Such a phenomenon is widely shared by
different DNNs for different tasks. This phenomenon indicates a cognition gap
between DNNs and human beings, and we call it a representation bottleneck. We
theoretically prove the underlying reason for the representation bottleneck.
Furthermore, we propose a loss to encourage/penalize the learning of
interactions of specific complexities, and analyze the representation
capacities of interactions of different complexities. |
doi_str_mv | 10.48550/arxiv.2111.06236 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2111_06236</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2111_06236</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-687a6f43492ee47e08b1539e8ab136e0c8eb181369f7b60eaddd586c8242b9f33</originalsourceid><addsrcrecordid>eNotz8tOwzAUBFBvWKDCB7DCP5DgV26cHdCWh1QVCXUfXSfXYBGcyLGq8vfQ0tXMbEY6jN1IURpbVeIO0yHsSyWlLAUoDZfsfhXmbtxTCvGDY-z5-jANGOJx5k_i7zQlmilmzGGM_HHMeaBI3RcfPV9tt_MVu_A4zHR9zgXbPa13y5di8_b8unzYFAg1FGBrBG-0aRSRqUlYJyvdkEUnNZDoLDlp_2rjaweCsO_7ykJnlVGu8Vov2O3_7YnQTil8Y_ppj5T2RNG__-hDhQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Discovering and Explaining the Representation Bottleneck of DNNs</title><source>arXiv.org</source><creator>Deng, Huiqi ; Ren, Qihan ; Zhang, Hao ; Zhang, Quanshi</creator><creatorcontrib>Deng, Huiqi ; Ren, Qihan ; Zhang, Hao ; Zhang, Quanshi</creatorcontrib><description>This paper explores the bottleneck of feature representations of deep neural
networks (DNNs), from the perspective of the complexity of interactions between
input variables encoded in DNNs. To this end, we focus on the multi-order
interaction between input variables, where the order represents the complexity
of interactions. We discover that a DNN is more likely to encode both too
simple interactions and too complex interactions, but usually fails to learn
interactions of intermediate complexity. Such a phenomenon is widely shared by
different DNNs for different tasks. This phenomenon indicates a cognition gap
between DNNs and human beings, and we call it a representation bottleneck. We
theoretically prove the underlying reason for the representation bottleneck.
Furthermore, we propose a loss to encourage/penalize the learning of
interactions of specific complexities, and analyze the representation
capacities of interactions of different complexities.</description><identifier>DOI: 10.48550/arxiv.2111.06236</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2021-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2111.06236$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2111.06236$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Deng, Huiqi</creatorcontrib><creatorcontrib>Ren, Qihan</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Zhang, Quanshi</creatorcontrib><title>Discovering and Explaining the Representation Bottleneck of DNNs</title><description>This paper explores the bottleneck of feature representations of deep neural
networks (DNNs), from the perspective of the complexity of interactions between
input variables encoded in DNNs. To this end, we focus on the multi-order
interaction between input variables, where the order represents the complexity
of interactions. We discover that a DNN is more likely to encode both too
simple interactions and too complex interactions, but usually fails to learn
interactions of intermediate complexity. Such a phenomenon is widely shared by
different DNNs for different tasks. This phenomenon indicates a cognition gap
between DNNs and human beings, and we call it a representation bottleneck. We
theoretically prove the underlying reason for the representation bottleneck.
Furthermore, we propose a loss to encourage/penalize the learning of
interactions of specific complexities, and analyze the representation
capacities of interactions of different complexities.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz8tOwzAUBFBvWKDCB7DCP5DgV26cHdCWh1QVCXUfXSfXYBGcyLGq8vfQ0tXMbEY6jN1IURpbVeIO0yHsSyWlLAUoDZfsfhXmbtxTCvGDY-z5-jANGOJx5k_i7zQlmilmzGGM_HHMeaBI3RcfPV9tt_MVu_A4zHR9zgXbPa13y5di8_b8unzYFAg1FGBrBG-0aRSRqUlYJyvdkEUnNZDoLDlp_2rjaweCsO_7ykJnlVGu8Vov2O3_7YnQTil8Y_ppj5T2RNG__-hDhQ</recordid><startdate>20211111</startdate><enddate>20211111</enddate><creator>Deng, Huiqi</creator><creator>Ren, Qihan</creator><creator>Zhang, Hao</creator><creator>Zhang, Quanshi</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20211111</creationdate><title>Discovering and Explaining the Representation Bottleneck of DNNs</title><author>Deng, Huiqi ; Ren, Qihan ; Zhang, Hao ; Zhang, Quanshi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-687a6f43492ee47e08b1539e8ab136e0c8eb181369f7b60eaddd586c8242b9f33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Deng, Huiqi</creatorcontrib><creatorcontrib>Ren, Qihan</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Zhang, Quanshi</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Deng, Huiqi</au><au>Ren, Qihan</au><au>Zhang, Hao</au><au>Zhang, Quanshi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Discovering and Explaining the Representation Bottleneck of DNNs</atitle><date>2021-11-11</date><risdate>2021</risdate><abstract>This paper explores the bottleneck of feature representations of deep neural
networks (DNNs), from the perspective of the complexity of interactions between
input variables encoded in DNNs. To this end, we focus on the multi-order
interaction between input variables, where the order represents the complexity
of interactions. We discover that a DNN is more likely to encode both too
simple interactions and too complex interactions, but usually fails to learn
interactions of intermediate complexity. Such a phenomenon is widely shared by
different DNNs for different tasks. This phenomenon indicates a cognition gap
between DNNs and human beings, and we call it a representation bottleneck. We
theoretically prove the underlying reason for the representation bottleneck.
Furthermore, we propose a loss to encourage/penalize the learning of
interactions of specific complexities, and analyze the representation
capacities of interactions of different complexities.</abstract><doi>10.48550/arxiv.2111.06236</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2111.06236 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2111_06236 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | Discovering and Explaining the Representation Bottleneck of DNNs |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T05%3A02%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Discovering%20and%20Explaining%20the%20Representation%20Bottleneck%20of%20DNNs&rft.au=Deng,%20Huiqi&rft.date=2021-11-11&rft_id=info:doi/10.48550/arxiv.2111.06236&rft_dat=%3Carxiv_GOX%3E2111_06236%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |