Deep learning via message passing algorithms based on belief propagation

Message-passing algorithms based on the Belief Propagation (BP) equations constitute a well-known distributed computational scheme. It is exact on tree-like graphical models and has also proven to be effective in many problems defined on graphs with loops (from inference to optimization, from signal...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2022-03
Hauptverfasser: Lucibello, Carlo, Pittorino, Fabrizio, Perugini, Gabriele, Zecchina, Riccardo
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Lucibello, Carlo
Pittorino, Fabrizio
Perugini, Gabriele
Zecchina, Riccardo
description Message-passing algorithms based on the Belief Propagation (BP) equations constitute a well-known distributed computational scheme. It is exact on tree-like graphical models and has also proven to be effective in many problems defined on graphs with loops (from inference to optimization, from signal processing to clustering). The BP-based scheme is fundamentally different from stochastic gradient descent (SGD), on which the current success of deep networks is based. In this paper, we present and adapt to mini-batch training on GPUs a family of BP-based message-passing algorithms with a reinforcement field that biases distributions towards locally entropic solutions. These algorithms are capable of training multi-layer neural networks with discrete weights and activations with performance comparable to SGD-inspired heuristics (BinaryNet) and are naturally well-adapted to continual learning. Furthermore, using these algorithms to estimate the marginals of the weights allows us to make approximate Bayesian predictions that have higher accuracy than point-wise solutions.
doi_str_mv 10.48550/arxiv.2110.14583
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2110_14583</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2587500345</sourcerecordid><originalsourceid>FETCH-LOGICAL-a525-6548112f3ed9dffec236978c88f5b8ce54f34d48385e7c866c8ac02ae734f09a3</originalsourceid><addsrcrecordid>eNotj8tqwzAQRUWh0JDmA7qqoGunsqSxJ8uSPlIIdJO9GdsjV8GxXSkJ7d_XeawuHC6Xe4R4SNXcIoB6pvDrj3OdjiC1gOZGTLQxaYJW6zsxi3GrlNJZrgHMRKxemQfZMoXOd408epI7jpEalgPFeGLUNn3w--9dlCVFrmXfyZJbz04OoR-oob3vu3tx66iNPLvmVGze3zbLVbL--vhcvqwTAg1JBhbTVDvD9aJ2jittskWOFaKDEisG64ytLRoEzivMsgqpUpo4N9apBZmpeLzMnjWLIfgdhb_ipFucdcfG06Uxnvs5cNwX2_4QuvFToQFzUMpYMP9HGVf0</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2587500345</pqid></control><display><type>article</type><title>Deep learning via message passing algorithms based on belief propagation</title><source>Freely Accessible Journals</source><source>arXiv.org</source><creator>Lucibello, Carlo ; Pittorino, Fabrizio ; Perugini, Gabriele ; Zecchina, Riccardo</creator><creatorcontrib>Lucibello, Carlo ; Pittorino, Fabrizio ; Perugini, Gabriele ; Zecchina, Riccardo</creatorcontrib><description>Message-passing algorithms based on the Belief Propagation (BP) equations constitute a well-known distributed computational scheme. It is exact on tree-like graphical models and has also proven to be effective in many problems defined on graphs with loops (from inference to optimization, from signal processing to clustering). The BP-based scheme is fundamentally different from stochastic gradient descent (SGD), on which the current success of deep networks is based. In this paper, we present and adapt to mini-batch training on GPUs a family of BP-based message-passing algorithms with a reinforcement field that biases distributions towards locally entropic solutions. These algorithms are capable of training multi-layer neural networks with discrete weights and activations with performance comparable to SGD-inspired heuristics (BinaryNet) and are naturally well-adapted to continual learning. Furthermore, using these algorithms to estimate the marginals of the weights allows us to make approximate Bayesian predictions that have higher accuracy than point-wise solutions.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2110.14583</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Clustering ; Computer Science - Learning ; Deep learning ; Machine learning ; Message passing ; Multilayers ; Neural networks ; Optimization ; Physics - Disordered Systems and Neural Networks ; Propagation ; Signal processing ; Statistics - Machine Learning ; Training</subject><ispartof>arXiv.org, 2022-03</ispartof><rights>2022. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2110.14583$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1088/2632-2153/ac7d3b$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Lucibello, Carlo</creatorcontrib><creatorcontrib>Pittorino, Fabrizio</creatorcontrib><creatorcontrib>Perugini, Gabriele</creatorcontrib><creatorcontrib>Zecchina, Riccardo</creatorcontrib><title>Deep learning via message passing algorithms based on belief propagation</title><title>arXiv.org</title><description>Message-passing algorithms based on the Belief Propagation (BP) equations constitute a well-known distributed computational scheme. It is exact on tree-like graphical models and has also proven to be effective in many problems defined on graphs with loops (from inference to optimization, from signal processing to clustering). The BP-based scheme is fundamentally different from stochastic gradient descent (SGD), on which the current success of deep networks is based. In this paper, we present and adapt to mini-batch training on GPUs a family of BP-based message-passing algorithms with a reinforcement field that biases distributions towards locally entropic solutions. These algorithms are capable of training multi-layer neural networks with discrete weights and activations with performance comparable to SGD-inspired heuristics (BinaryNet) and are naturally well-adapted to continual learning. Furthermore, using these algorithms to estimate the marginals of the weights allows us to make approximate Bayesian predictions that have higher accuracy than point-wise solutions.</description><subject>Algorithms</subject><subject>Clustering</subject><subject>Computer Science - Learning</subject><subject>Deep learning</subject><subject>Machine learning</subject><subject>Message passing</subject><subject>Multilayers</subject><subject>Neural networks</subject><subject>Optimization</subject><subject>Physics - Disordered Systems and Neural Networks</subject><subject>Propagation</subject><subject>Signal processing</subject><subject>Statistics - Machine Learning</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotj8tqwzAQRUWh0JDmA7qqoGunsqSxJ8uSPlIIdJO9GdsjV8GxXSkJ7d_XeawuHC6Xe4R4SNXcIoB6pvDrj3OdjiC1gOZGTLQxaYJW6zsxi3GrlNJZrgHMRKxemQfZMoXOd408epI7jpEalgPFeGLUNn3w--9dlCVFrmXfyZJbz04OoR-oob3vu3tx66iNPLvmVGze3zbLVbL--vhcvqwTAg1JBhbTVDvD9aJ2jittskWOFaKDEisG64ytLRoEzivMsgqpUpo4N9apBZmpeLzMnjWLIfgdhb_ipFucdcfG06Uxnvs5cNwX2_4QuvFToQFzUMpYMP9HGVf0</recordid><startdate>20220315</startdate><enddate>20220315</enddate><creator>Lucibello, Carlo</creator><creator>Pittorino, Fabrizio</creator><creator>Perugini, Gabriele</creator><creator>Zecchina, Riccardo</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20220315</creationdate><title>Deep learning via message passing algorithms based on belief propagation</title><author>Lucibello, Carlo ; Pittorino, Fabrizio ; Perugini, Gabriele ; Zecchina, Riccardo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a525-6548112f3ed9dffec236978c88f5b8ce54f34d48385e7c866c8ac02ae734f09a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Clustering</topic><topic>Computer Science - Learning</topic><topic>Deep learning</topic><topic>Machine learning</topic><topic>Message passing</topic><topic>Multilayers</topic><topic>Neural networks</topic><topic>Optimization</topic><topic>Physics - Disordered Systems and Neural Networks</topic><topic>Propagation</topic><topic>Signal processing</topic><topic>Statistics - Machine Learning</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Lucibello, Carlo</creatorcontrib><creatorcontrib>Pittorino, Fabrizio</creatorcontrib><creatorcontrib>Perugini, Gabriele</creatorcontrib><creatorcontrib>Zecchina, Riccardo</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Lucibello, Carlo</au><au>Pittorino, Fabrizio</au><au>Perugini, Gabriele</au><au>Zecchina, Riccardo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep learning via message passing algorithms based on belief propagation</atitle><jtitle>arXiv.org</jtitle><date>2022-03-15</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>Message-passing algorithms based on the Belief Propagation (BP) equations constitute a well-known distributed computational scheme. It is exact on tree-like graphical models and has also proven to be effective in many problems defined on graphs with loops (from inference to optimization, from signal processing to clustering). The BP-based scheme is fundamentally different from stochastic gradient descent (SGD), on which the current success of deep networks is based. In this paper, we present and adapt to mini-batch training on GPUs a family of BP-based message-passing algorithms with a reinforcement field that biases distributions towards locally entropic solutions. These algorithms are capable of training multi-layer neural networks with discrete weights and activations with performance comparable to SGD-inspired heuristics (BinaryNet) and are naturally well-adapted to continual learning. Furthermore, using these algorithms to estimate the marginals of the weights allows us to make approximate Bayesian predictions that have higher accuracy than point-wise solutions.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2110.14583</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-03
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2110_14583
source Freely Accessible Journals; arXiv.org
subjects Algorithms
Clustering
Computer Science - Learning
Deep learning
Machine learning
Message passing
Multilayers
Neural networks
Optimization
Physics - Disordered Systems and Neural Networks
Propagation
Signal processing
Statistics - Machine Learning
Training
title Deep learning via message passing algorithms based on belief propagation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T19%3A23%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20learning%20via%20message%20passing%20algorithms%20based%20on%20belief%20propagation&rft.jtitle=arXiv.org&rft.au=Lucibello,%20Carlo&rft.date=2022-03-15&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2110.14583&rft_dat=%3Cproquest_arxiv%3E2587500345%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2587500345&rft_id=info:pmid/&rfr_iscdi=true