Constructing Deep Neural Networks by Bayesian Network Structure Learning

We introduce a principled approach for unsupervised structure learning of deep neural networks. We propose a new interpretation for depth and inter-layer connectivity where conditional independencies in the input distribution are encoded hierarchically in the network structure. Thus, the depth of th...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Rohekar, Raanan Y, Nisimov, Shami, Gurwicz, Yaniv, Koren, Guy, Novik, Gal
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Rohekar, Raanan Y
Nisimov, Shami
Gurwicz, Yaniv
Koren, Guy
Novik, Gal
description We introduce a principled approach for unsupervised structure learning of deep neural networks. We propose a new interpretation for depth and inter-layer connectivity where conditional independencies in the input distribution are encoded hierarchically in the network structure. Thus, the depth of the network is determined inherently. The proposed method casts the problem of neural network structure learning as a problem of Bayesian network structure learning. Then, instead of directly learning the discriminative structure, it learns a generative graph, constructs its stochastic inverse, and then constructs a discriminative graph. We prove that conditional-dependency relations among the latent variables in the generative graph are preserved in the class-conditional discriminative graph. We demonstrate on image classification benchmarks that the deepest layers (convolutional and dense) of common networks can be replaced by significantly smaller learned structures, while maintaining classification accuracy---state-of-the-art on tested benchmarks. Our structure learning algorithm requires a small computational cost and runs efficiently on a standard desktop CPU.
doi_str_mv 10.48550/arxiv.1806.09141
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1806_09141</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1806_09141</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-b541c719d42218ca72f28029a2f1dab5db8c1b1e18681e5a07c806358cb7df963</originalsourceid><addsrcrecordid>eNo1j8tugzAURL3Jokr6AV3VPwD1NRibZUqbh4TaRbNH1-ZSoSYkMtCEvy-hyepII81oDmNPIMLYKCVe0F_q3xCMSEKRQgwPbJMdm7bzvevq5pu_EZ34B_Ue9yO689H_tNwO_BUHamts7iH_miq9J54T-mbsLtiswn1LjzfO2W71vss2Qf653mbLPMBEQ2BVDE5DWsZSgnGoZSWNkCnKCkq0qrTGgQUCkxgghUK78WykjLO6rNIkmrPn_9lJpTj5-oB-KK5KxaQU_QHIpEbK</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Constructing Deep Neural Networks by Bayesian Network Structure Learning</title><source>arXiv.org</source><creator>Rohekar, Raanan Y ; Nisimov, Shami ; Gurwicz, Yaniv ; Koren, Guy ; Novik, Gal</creator><creatorcontrib>Rohekar, Raanan Y ; Nisimov, Shami ; Gurwicz, Yaniv ; Koren, Guy ; Novik, Gal</creatorcontrib><description>We introduce a principled approach for unsupervised structure learning of deep neural networks. We propose a new interpretation for depth and inter-layer connectivity where conditional independencies in the input distribution are encoded hierarchically in the network structure. Thus, the depth of the network is determined inherently. The proposed method casts the problem of neural network structure learning as a problem of Bayesian network structure learning. Then, instead of directly learning the discriminative structure, it learns a generative graph, constructs its stochastic inverse, and then constructs a discriminative graph. We prove that conditional-dependency relations among the latent variables in the generative graph are preserved in the class-conditional discriminative graph. We demonstrate on image classification benchmarks that the deepest layers (convolutional and dense) of common networks can be replaced by significantly smaller learned structures, while maintaining classification accuracy---state-of-the-art on tested benchmarks. Our structure learning algorithm requires a small computational cost and runs efficiently on a standard desktop CPU.</description><identifier>DOI: 10.48550/arxiv.1806.09141</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2018-06</creationdate><rights>http://creativecommons.org/licenses/by-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1806.09141$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1806.09141$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Rohekar, Raanan Y</creatorcontrib><creatorcontrib>Nisimov, Shami</creatorcontrib><creatorcontrib>Gurwicz, Yaniv</creatorcontrib><creatorcontrib>Koren, Guy</creatorcontrib><creatorcontrib>Novik, Gal</creatorcontrib><title>Constructing Deep Neural Networks by Bayesian Network Structure Learning</title><description>We introduce a principled approach for unsupervised structure learning of deep neural networks. We propose a new interpretation for depth and inter-layer connectivity where conditional independencies in the input distribution are encoded hierarchically in the network structure. Thus, the depth of the network is determined inherently. The proposed method casts the problem of neural network structure learning as a problem of Bayesian network structure learning. Then, instead of directly learning the discriminative structure, it learns a generative graph, constructs its stochastic inverse, and then constructs a discriminative graph. We prove that conditional-dependency relations among the latent variables in the generative graph are preserved in the class-conditional discriminative graph. We demonstrate on image classification benchmarks that the deepest layers (convolutional and dense) of common networks can be replaced by significantly smaller learned structures, while maintaining classification accuracy---state-of-the-art on tested benchmarks. Our structure learning algorithm requires a small computational cost and runs efficiently on a standard desktop CPU.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNo1j8tugzAURL3Jokr6AV3VPwD1NRibZUqbh4TaRbNH1-ZSoSYkMtCEvy-hyepII81oDmNPIMLYKCVe0F_q3xCMSEKRQgwPbJMdm7bzvevq5pu_EZ34B_Ue9yO689H_tNwO_BUHamts7iH_miq9J54T-mbsLtiswn1LjzfO2W71vss2Qf653mbLPMBEQ2BVDE5DWsZSgnGoZSWNkCnKCkq0qrTGgQUCkxgghUK78WykjLO6rNIkmrPn_9lJpTj5-oB-KK5KxaQU_QHIpEbK</recordid><startdate>20180624</startdate><enddate>20180624</enddate><creator>Rohekar, Raanan Y</creator><creator>Nisimov, Shami</creator><creator>Gurwicz, Yaniv</creator><creator>Koren, Guy</creator><creator>Novik, Gal</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20180624</creationdate><title>Constructing Deep Neural Networks by Bayesian Network Structure Learning</title><author>Rohekar, Raanan Y ; Nisimov, Shami ; Gurwicz, Yaniv ; Koren, Guy ; Novik, Gal</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-b541c719d42218ca72f28029a2f1dab5db8c1b1e18681e5a07c806358cb7df963</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Rohekar, Raanan Y</creatorcontrib><creatorcontrib>Nisimov, Shami</creatorcontrib><creatorcontrib>Gurwicz, Yaniv</creatorcontrib><creatorcontrib>Koren, Guy</creatorcontrib><creatorcontrib>Novik, Gal</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Rohekar, Raanan Y</au><au>Nisimov, Shami</au><au>Gurwicz, Yaniv</au><au>Koren, Guy</au><au>Novik, Gal</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Constructing Deep Neural Networks by Bayesian Network Structure Learning</atitle><date>2018-06-24</date><risdate>2018</risdate><abstract>We introduce a principled approach for unsupervised structure learning of deep neural networks. We propose a new interpretation for depth and inter-layer connectivity where conditional independencies in the input distribution are encoded hierarchically in the network structure. Thus, the depth of the network is determined inherently. The proposed method casts the problem of neural network structure learning as a problem of Bayesian network structure learning. Then, instead of directly learning the discriminative structure, it learns a generative graph, constructs its stochastic inverse, and then constructs a discriminative graph. We prove that conditional-dependency relations among the latent variables in the generative graph are preserved in the class-conditional discriminative graph. We demonstrate on image classification benchmarks that the deepest layers (convolutional and dense) of common networks can be replaced by significantly smaller learned structures, while maintaining classification accuracy---state-of-the-art on tested benchmarks. Our structure learning algorithm requires a small computational cost and runs efficiently on a standard desktop CPU.</abstract><doi>10.48550/arxiv.1806.09141</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1806.09141
ispartof
issn
language eng
recordid cdi_arxiv_primary_1806_09141
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
Statistics - Machine Learning
title Constructing Deep Neural Networks by Bayesian Network Structure Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-21T15%3A09%3A15IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Constructing%20Deep%20Neural%20Networks%20by%20Bayesian%20Network%20Structure%20Learning&rft.au=Rohekar,%20Raanan%20Y&rft.date=2018-06-24&rft_id=info:doi/10.48550/arxiv.1806.09141&rft_dat=%3Carxiv_GOX%3E1806_09141%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true