Constructing Deep Neural Networks by Bayesian Network Structure Learning
We introduce a principled approach for unsupervised structure learning of deep neural networks. We propose a new interpretation for depth and inter-layer connectivity where conditional independencies in the input distribution are encoded hierarchically in the network structure. Thus, the depth of th...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Rohekar, Raanan Y Nisimov, Shami Gurwicz, Yaniv Koren, Guy Novik, Gal |
description | We introduce a principled approach for unsupervised structure learning of
deep neural networks. We propose a new interpretation for depth and inter-layer
connectivity where conditional independencies in the input distribution are
encoded hierarchically in the network structure. Thus, the depth of the network
is determined inherently. The proposed method casts the problem of neural
network structure learning as a problem of Bayesian network structure learning.
Then, instead of directly learning the discriminative structure, it learns a
generative graph, constructs its stochastic inverse, and then constructs a
discriminative graph. We prove that conditional-dependency relations among the
latent variables in the generative graph are preserved in the class-conditional
discriminative graph. We demonstrate on image classification benchmarks that
the deepest layers (convolutional and dense) of common networks can be replaced
by significantly smaller learned structures, while maintaining classification
accuracy---state-of-the-art on tested benchmarks. Our structure learning
algorithm requires a small computational cost and runs efficiently on a
standard desktop CPU. |
doi_str_mv | 10.48550/arxiv.1806.09141 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1806_09141</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1806_09141</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-b541c719d42218ca72f28029a2f1dab5db8c1b1e18681e5a07c806358cb7df963</originalsourceid><addsrcrecordid>eNo1j8tugzAURL3Jokr6AV3VPwD1NRibZUqbh4TaRbNH1-ZSoSYkMtCEvy-hyepII81oDmNPIMLYKCVe0F_q3xCMSEKRQgwPbJMdm7bzvevq5pu_EZ34B_Ue9yO689H_tNwO_BUHamts7iH_miq9J54T-mbsLtiswn1LjzfO2W71vss2Qf653mbLPMBEQ2BVDE5DWsZSgnGoZSWNkCnKCkq0qrTGgQUCkxgghUK78WykjLO6rNIkmrPn_9lJpTj5-oB-KK5KxaQU_QHIpEbK</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Constructing Deep Neural Networks by Bayesian Network Structure Learning</title><source>arXiv.org</source><creator>Rohekar, Raanan Y ; Nisimov, Shami ; Gurwicz, Yaniv ; Koren, Guy ; Novik, Gal</creator><creatorcontrib>Rohekar, Raanan Y ; Nisimov, Shami ; Gurwicz, Yaniv ; Koren, Guy ; Novik, Gal</creatorcontrib><description>We introduce a principled approach for unsupervised structure learning of
deep neural networks. We propose a new interpretation for depth and inter-layer
connectivity where conditional independencies in the input distribution are
encoded hierarchically in the network structure. Thus, the depth of the network
is determined inherently. The proposed method casts the problem of neural
network structure learning as a problem of Bayesian network structure learning.
Then, instead of directly learning the discriminative structure, it learns a
generative graph, constructs its stochastic inverse, and then constructs a
discriminative graph. We prove that conditional-dependency relations among the
latent variables in the generative graph are preserved in the class-conditional
discriminative graph. We demonstrate on image classification benchmarks that
the deepest layers (convolutional and dense) of common networks can be replaced
by significantly smaller learned structures, while maintaining classification
accuracy---state-of-the-art on tested benchmarks. Our structure learning
algorithm requires a small computational cost and runs efficiently on a
standard desktop CPU.</description><identifier>DOI: 10.48550/arxiv.1806.09141</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2018-06</creationdate><rights>http://creativecommons.org/licenses/by-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1806.09141$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1806.09141$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Rohekar, Raanan Y</creatorcontrib><creatorcontrib>Nisimov, Shami</creatorcontrib><creatorcontrib>Gurwicz, Yaniv</creatorcontrib><creatorcontrib>Koren, Guy</creatorcontrib><creatorcontrib>Novik, Gal</creatorcontrib><title>Constructing Deep Neural Networks by Bayesian Network Structure Learning</title><description>We introduce a principled approach for unsupervised structure learning of
deep neural networks. We propose a new interpretation for depth and inter-layer
connectivity where conditional independencies in the input distribution are
encoded hierarchically in the network structure. Thus, the depth of the network
is determined inherently. The proposed method casts the problem of neural
network structure learning as a problem of Bayesian network structure learning.
Then, instead of directly learning the discriminative structure, it learns a
generative graph, constructs its stochastic inverse, and then constructs a
discriminative graph. We prove that conditional-dependency relations among the
latent variables in the generative graph are preserved in the class-conditional
discriminative graph. We demonstrate on image classification benchmarks that
the deepest layers (convolutional and dense) of common networks can be replaced
by significantly smaller learned structures, while maintaining classification
accuracy---state-of-the-art on tested benchmarks. Our structure learning
algorithm requires a small computational cost and runs efficiently on a
standard desktop CPU.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNo1j8tugzAURL3Jokr6AV3VPwD1NRibZUqbh4TaRbNH1-ZSoSYkMtCEvy-hyepII81oDmNPIMLYKCVe0F_q3xCMSEKRQgwPbJMdm7bzvevq5pu_EZ34B_Ue9yO689H_tNwO_BUHamts7iH_miq9J54T-mbsLtiswn1LjzfO2W71vss2Qf653mbLPMBEQ2BVDE5DWsZSgnGoZSWNkCnKCkq0qrTGgQUCkxgghUK78WykjLO6rNIkmrPn_9lJpTj5-oB-KK5KxaQU_QHIpEbK</recordid><startdate>20180624</startdate><enddate>20180624</enddate><creator>Rohekar, Raanan Y</creator><creator>Nisimov, Shami</creator><creator>Gurwicz, Yaniv</creator><creator>Koren, Guy</creator><creator>Novik, Gal</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20180624</creationdate><title>Constructing Deep Neural Networks by Bayesian Network Structure Learning</title><author>Rohekar, Raanan Y ; Nisimov, Shami ; Gurwicz, Yaniv ; Koren, Guy ; Novik, Gal</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-b541c719d42218ca72f28029a2f1dab5db8c1b1e18681e5a07c806358cb7df963</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Rohekar, Raanan Y</creatorcontrib><creatorcontrib>Nisimov, Shami</creatorcontrib><creatorcontrib>Gurwicz, Yaniv</creatorcontrib><creatorcontrib>Koren, Guy</creatorcontrib><creatorcontrib>Novik, Gal</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Rohekar, Raanan Y</au><au>Nisimov, Shami</au><au>Gurwicz, Yaniv</au><au>Koren, Guy</au><au>Novik, Gal</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Constructing Deep Neural Networks by Bayesian Network Structure Learning</atitle><date>2018-06-24</date><risdate>2018</risdate><abstract>We introduce a principled approach for unsupervised structure learning of
deep neural networks. We propose a new interpretation for depth and inter-layer
connectivity where conditional independencies in the input distribution are
encoded hierarchically in the network structure. Thus, the depth of the network
is determined inherently. The proposed method casts the problem of neural
network structure learning as a problem of Bayesian network structure learning.
Then, instead of directly learning the discriminative structure, it learns a
generative graph, constructs its stochastic inverse, and then constructs a
discriminative graph. We prove that conditional-dependency relations among the
latent variables in the generative graph are preserved in the class-conditional
discriminative graph. We demonstrate on image classification benchmarks that
the deepest layers (convolutional and dense) of common networks can be replaced
by significantly smaller learned structures, while maintaining classification
accuracy---state-of-the-art on tested benchmarks. Our structure learning
algorithm requires a small computational cost and runs efficiently on a
standard desktop CPU.</abstract><doi>10.48550/arxiv.1806.09141</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1806.09141 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1806_09141 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning Statistics - Machine Learning |
title | Constructing Deep Neural Networks by Bayesian Network Structure Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-21T15%3A09%3A15IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Constructing%20Deep%20Neural%20Networks%20by%20Bayesian%20Network%20Structure%20Learning&rft.au=Rohekar,%20Raanan%20Y&rft.date=2018-06-24&rft_id=info:doi/10.48550/arxiv.1806.09141&rft_dat=%3Carxiv_GOX%3E1806_09141%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |