Residual-based attention and connection to information bottleneck theory in PINNs

Driven by the need for more efficient and seamless integration of physical models and data, physics-informed neural networks (PINNs) have seen a surge of interest in recent years. However, ensuring the reliability of their convergence and accuracy remains a challenge. In this work, we propose an eff...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2023-07
Hauptverfasser: Anagnostopoulos, Sokratis J, Juan Diego Toscano, Stergiopulos, Nikolaos, Karniadakis, George Em
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Anagnostopoulos, Sokratis J
Juan Diego Toscano
Stergiopulos, Nikolaos
Karniadakis, George Em
description Driven by the need for more efficient and seamless integration of physical models and data, physics-informed neural networks (PINNs) have seen a surge of interest in recent years. However, ensuring the reliability of their convergence and accuracy remains a challenge. In this work, we propose an efficient, gradient-less weighting scheme for PINNs, that accelerates the convergence of dynamic or static systems. This simple yet effective attention mechanism is a function of the evolving cumulative residuals and aims to make the optimizer aware of problematic regions at no extra computational cost or adversarial learning. We illustrate that this general method consistently achieves a relative \(L^{2}\) error of the order of \(10^{-5}\) using standard optimizers on typical benchmark cases of the literature. Furthermore, by investigating the evolution of weights during training, we identify two distinct learning phases reminiscent of the fitting and diffusion phases proposed by the information bottleneck (IB) theory. Subsequent gradient analysis supports this hypothesis by aligning the transition from high to low signal-to-noise ratio (SNR) with the transition from fitting to diffusion regimes of the adopted weights. This novel correlation between PINNs and IB theory could open future possibilities for understanding the underlying mechanisms behind the training and stability of PINNs and, more broadly, of neural operators.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2832892187</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2832892187</sourcerecordid><originalsourceid>FETCH-proquest_journals_28328921873</originalsourceid><addsrcrecordid>eNqNysEKgkAUheEhCJLyHQZaC3onc1pHURupaC-jXkmzuTVzXfT2SfQArQ4_35mIAJRKIr0CmInQ-y6OY1hnkKYqEOcL-rYeTB-VxmMtDTNabslKY2tZkbVYfZNJtrYh9zDfLIm5xxHvkm9I7j2qPB3z3C_EtDG9x_C3c7Hc767bQ_R09BrQc9HR4OxIBWgFegOJztR_rw_rQz_G</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2832892187</pqid></control><display><type>article</type><title>Residual-based attention and connection to information bottleneck theory in PINNs</title><source>Freely Accessible Journals</source><creator>Anagnostopoulos, Sokratis J ; Juan Diego Toscano ; Stergiopulos, Nikolaos ; Karniadakis, George Em</creator><creatorcontrib>Anagnostopoulos, Sokratis J ; Juan Diego Toscano ; Stergiopulos, Nikolaos ; Karniadakis, George Em</creatorcontrib><description>Driven by the need for more efficient and seamless integration of physical models and data, physics-informed neural networks (PINNs) have seen a surge of interest in recent years. However, ensuring the reliability of their convergence and accuracy remains a challenge. In this work, we propose an efficient, gradient-less weighting scheme for PINNs, that accelerates the convergence of dynamic or static systems. This simple yet effective attention mechanism is a function of the evolving cumulative residuals and aims to make the optimizer aware of problematic regions at no extra computational cost or adversarial learning. We illustrate that this general method consistently achieves a relative \(L^{2}\) error of the order of \(10^{-5}\) using standard optimizers on typical benchmark cases of the literature. Furthermore, by investigating the evolution of weights during training, we identify two distinct learning phases reminiscent of the fitting and diffusion phases proposed by the information bottleneck (IB) theory. Subsequent gradient analysis supports this hypothesis by aligning the transition from high to low signal-to-noise ratio (SNR) with the transition from fitting to diffusion regimes of the adopted weights. This novel correlation between PINNs and IB theory could open future possibilities for understanding the underlying mechanisms behind the training and stability of PINNs and, more broadly, of neural operators.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Convergence ; Learning ; Neural networks ; Signal to noise ratio ; System effectiveness ; Training</subject><ispartof>arXiv.org, 2023-07</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>781,785</link.rule.ids></links><search><creatorcontrib>Anagnostopoulos, Sokratis J</creatorcontrib><creatorcontrib>Juan Diego Toscano</creatorcontrib><creatorcontrib>Stergiopulos, Nikolaos</creatorcontrib><creatorcontrib>Karniadakis, George Em</creatorcontrib><title>Residual-based attention and connection to information bottleneck theory in PINNs</title><title>arXiv.org</title><description>Driven by the need for more efficient and seamless integration of physical models and data, physics-informed neural networks (PINNs) have seen a surge of interest in recent years. However, ensuring the reliability of their convergence and accuracy remains a challenge. In this work, we propose an efficient, gradient-less weighting scheme for PINNs, that accelerates the convergence of dynamic or static systems. This simple yet effective attention mechanism is a function of the evolving cumulative residuals and aims to make the optimizer aware of problematic regions at no extra computational cost or adversarial learning. We illustrate that this general method consistently achieves a relative \(L^{2}\) error of the order of \(10^{-5}\) using standard optimizers on typical benchmark cases of the literature. Furthermore, by investigating the evolution of weights during training, we identify two distinct learning phases reminiscent of the fitting and diffusion phases proposed by the information bottleneck (IB) theory. Subsequent gradient analysis supports this hypothesis by aligning the transition from high to low signal-to-noise ratio (SNR) with the transition from fitting to diffusion regimes of the adopted weights. This novel correlation between PINNs and IB theory could open future possibilities for understanding the underlying mechanisms behind the training and stability of PINNs and, more broadly, of neural operators.</description><subject>Convergence</subject><subject>Learning</subject><subject>Neural networks</subject><subject>Signal to noise ratio</subject><subject>System effectiveness</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNysEKgkAUheEhCJLyHQZaC3onc1pHURupaC-jXkmzuTVzXfT2SfQArQ4_35mIAJRKIr0CmInQ-y6OY1hnkKYqEOcL-rYeTB-VxmMtDTNabslKY2tZkbVYfZNJtrYh9zDfLIm5xxHvkm9I7j2qPB3z3C_EtDG9x_C3c7Hc767bQ_R09BrQc9HR4OxIBWgFegOJztR_rw_rQz_G</recordid><startdate>20230701</startdate><enddate>20230701</enddate><creator>Anagnostopoulos, Sokratis J</creator><creator>Juan Diego Toscano</creator><creator>Stergiopulos, Nikolaos</creator><creator>Karniadakis, George Em</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20230701</creationdate><title>Residual-based attention and connection to information bottleneck theory in PINNs</title><author>Anagnostopoulos, Sokratis J ; Juan Diego Toscano ; Stergiopulos, Nikolaos ; Karniadakis, George Em</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_28328921873</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Convergence</topic><topic>Learning</topic><topic>Neural networks</topic><topic>Signal to noise ratio</topic><topic>System effectiveness</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Anagnostopoulos, Sokratis J</creatorcontrib><creatorcontrib>Juan Diego Toscano</creatorcontrib><creatorcontrib>Stergiopulos, Nikolaos</creatorcontrib><creatorcontrib>Karniadakis, George Em</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Anagnostopoulos, Sokratis J</au><au>Juan Diego Toscano</au><au>Stergiopulos, Nikolaos</au><au>Karniadakis, George Em</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Residual-based attention and connection to information bottleneck theory in PINNs</atitle><jtitle>arXiv.org</jtitle><date>2023-07-01</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Driven by the need for more efficient and seamless integration of physical models and data, physics-informed neural networks (PINNs) have seen a surge of interest in recent years. However, ensuring the reliability of their convergence and accuracy remains a challenge. In this work, we propose an efficient, gradient-less weighting scheme for PINNs, that accelerates the convergence of dynamic or static systems. This simple yet effective attention mechanism is a function of the evolving cumulative residuals and aims to make the optimizer aware of problematic regions at no extra computational cost or adversarial learning. We illustrate that this general method consistently achieves a relative \(L^{2}\) error of the order of \(10^{-5}\) using standard optimizers on typical benchmark cases of the literature. Furthermore, by investigating the evolution of weights during training, we identify two distinct learning phases reminiscent of the fitting and diffusion phases proposed by the information bottleneck (IB) theory. Subsequent gradient analysis supports this hypothesis by aligning the transition from high to low signal-to-noise ratio (SNR) with the transition from fitting to diffusion regimes of the adopted weights. This novel correlation between PINNs and IB theory could open future possibilities for understanding the underlying mechanisms behind the training and stability of PINNs and, more broadly, of neural operators.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-07
issn 2331-8422
language eng
recordid cdi_proquest_journals_2832892187
source Freely Accessible Journals
subjects Convergence
Learning
Neural networks
Signal to noise ratio
System effectiveness
Training
title Residual-based attention and connection to information bottleneck theory in PINNs
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-14T05%3A04%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Residual-based%20attention%20and%20connection%20to%20information%20bottleneck%20theory%20in%20PINNs&rft.jtitle=arXiv.org&rft.au=Anagnostopoulos,%20Sokratis%20J&rft.date=2023-07-01&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2832892187%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2832892187&rft_id=info:pmid/&rfr_iscdi=true