Reliable extrapolation of deep neural operators informed by physics or sparse observations

Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks. As promising surrogate solvers of partial differential equations (PDEs) for real-time prediction, deep neural operators such as deep operator networks (DeepONets) provide a new s...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computer methods in applied mechanics and engineering 2023-05, Vol.412
Hauptverfasser: Zhu, Min, Zhang, Handi, Jiao, Anran, Karniadakis, George Em, Lu, Lu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title Computer methods in applied mechanics and engineering
container_volume 412
creator Zhu, Min
Zhang, Handi
Jiao, Anran
Karniadakis, George Em
Lu, Lu
description Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks. As promising surrogate solvers of partial differential equations (PDEs) for real-time prediction, deep neural operators such as deep operator networks (DeepONets) provide a new simulation paradigm in science and engineering. Pure data-driven neural operators and deep learning models, in general, are usually limited to interpolation scenarios, where new predictions utilize inputs within the support of the training set. However, in the inference stage of real-world applications, the input may lie outside the support, i.e., extrapolation is required, which may result to large errors and unavoidable failure of deep learning models. Here, we address this challenge of extrapolation for deep neural operators. First, we systematically investigate the extrapolation behavior of DeepONets by quantifying the extrapolation complexity, via the 2-Wasserstein distance between two function spaces and propose a new strategy of bias–variance trade-off for extrapolation with respect to model capacity. Subsequently, we develop a complete workflow, including extrapolation determination, and we propose five reliable learning methods that guarantee a safe prediction under extrapolation by requiring additional information—the governing PDEs of the system or sparse new observations. The proposed methods are based on either fine-tuning a pre-trained DeepONet or multifidelity learning. We demonstrate the effectiveness of the proposed framework for various types of parametric PDEs. Furthermore, our systematic comparisons provide practical guidelines for selecting a proper extrapolation method depending on the available information, desired accuracy, and required inference speed.
format Article
fullrecord <record><control><sourceid>osti</sourceid><recordid>TN_cdi_osti_scitechconnect_1991293</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1991293</sourcerecordid><originalsourceid>FETCH-osti_scitechconnect_19912933</originalsourceid><addsrcrecordid>eNqNjMsKwjAQAIMoWB__sHgvNK2l7VkUz-LJi6TplkZiNmSj2L9XxA9wLnMZZiISWVdNmsuinooky7ZlWtV5ORcL5lv2oZZ5Ii4ntEa1FgFfMShPVkVDDqiHDtGDw0dQFshjUJECg3E9hTt20I7gh5GNZqAA7FVgBGoZw_O74JWY9coyrn9eis1hf94dU-JorqxNRD1ocg51vMqmkXlTFH9Fb1a2RRs</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Reliable extrapolation of deep neural operators informed by physics or sparse observations</title><source>Elsevier ScienceDirect Journals</source><creator>Zhu, Min ; Zhang, Handi ; Jiao, Anran ; Karniadakis, George Em ; Lu, Lu</creator><creatorcontrib>Zhu, Min ; Zhang, Handi ; Jiao, Anran ; Karniadakis, George Em ; Lu, Lu ; Univ. of Pennsylvania, Philadelphia, PA (United States)</creatorcontrib><description>Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks. As promising surrogate solvers of partial differential equations (PDEs) for real-time prediction, deep neural operators such as deep operator networks (DeepONets) provide a new simulation paradigm in science and engineering. Pure data-driven neural operators and deep learning models, in general, are usually limited to interpolation scenarios, where new predictions utilize inputs within the support of the training set. However, in the inference stage of real-world applications, the input may lie outside the support, i.e., extrapolation is required, which may result to large errors and unavoidable failure of deep learning models. Here, we address this challenge of extrapolation for deep neural operators. First, we systematically investigate the extrapolation behavior of DeepONets by quantifying the extrapolation complexity, via the 2-Wasserstein distance between two function spaces and propose a new strategy of bias–variance trade-off for extrapolation with respect to model capacity. Subsequently, we develop a complete workflow, including extrapolation determination, and we propose five reliable learning methods that guarantee a safe prediction under extrapolation by requiring additional information—the governing PDEs of the system or sparse new observations. The proposed methods are based on either fine-tuning a pre-trained DeepONet or multifidelity learning. We demonstrate the effectiveness of the proposed framework for various types of parametric PDEs. Furthermore, our systematic comparisons provide practical guidelines for selecting a proper extrapolation method depending on the available information, desired accuracy, and required inference speed.</description><identifier>ISSN: 0045-7825</identifier><identifier>EISSN: 1879-2138</identifier><language>eng</language><publisher>United States: Elsevier</publisher><subject>DeepONet ; ENGINEERING ; Extrapolation complexity ; Fine-tuning ; Multifidelity learning ; Neural operators ; Out-of-distribution inference</subject><ispartof>Computer methods in applied mechanics and engineering, 2023-05, Vol.412</ispartof><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>0000000254765768 ; 0000000345505662</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,776,780,881</link.rule.ids><backlink>$$Uhttps://www.osti.gov/servlets/purl/1991293$$D View this record in Osti.gov$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhu, Min</creatorcontrib><creatorcontrib>Zhang, Handi</creatorcontrib><creatorcontrib>Jiao, Anran</creatorcontrib><creatorcontrib>Karniadakis, George Em</creatorcontrib><creatorcontrib>Lu, Lu</creatorcontrib><creatorcontrib>Univ. of Pennsylvania, Philadelphia, PA (United States)</creatorcontrib><title>Reliable extrapolation of deep neural operators informed by physics or sparse observations</title><title>Computer methods in applied mechanics and engineering</title><description>Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks. As promising surrogate solvers of partial differential equations (PDEs) for real-time prediction, deep neural operators such as deep operator networks (DeepONets) provide a new simulation paradigm in science and engineering. Pure data-driven neural operators and deep learning models, in general, are usually limited to interpolation scenarios, where new predictions utilize inputs within the support of the training set. However, in the inference stage of real-world applications, the input may lie outside the support, i.e., extrapolation is required, which may result to large errors and unavoidable failure of deep learning models. Here, we address this challenge of extrapolation for deep neural operators. First, we systematically investigate the extrapolation behavior of DeepONets by quantifying the extrapolation complexity, via the 2-Wasserstein distance between two function spaces and propose a new strategy of bias–variance trade-off for extrapolation with respect to model capacity. Subsequently, we develop a complete workflow, including extrapolation determination, and we propose five reliable learning methods that guarantee a safe prediction under extrapolation by requiring additional information—the governing PDEs of the system or sparse new observations. The proposed methods are based on either fine-tuning a pre-trained DeepONet or multifidelity learning. We demonstrate the effectiveness of the proposed framework for various types of parametric PDEs. Furthermore, our systematic comparisons provide practical guidelines for selecting a proper extrapolation method depending on the available information, desired accuracy, and required inference speed.</description><subject>DeepONet</subject><subject>ENGINEERING</subject><subject>Extrapolation complexity</subject><subject>Fine-tuning</subject><subject>Multifidelity learning</subject><subject>Neural operators</subject><subject>Out-of-distribution inference</subject><issn>0045-7825</issn><issn>1879-2138</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNqNjMsKwjAQAIMoWB__sHgvNK2l7VkUz-LJi6TplkZiNmSj2L9XxA9wLnMZZiISWVdNmsuinooky7ZlWtV5ORcL5lv2oZZ5Ii4ntEa1FgFfMShPVkVDDqiHDtGDw0dQFshjUJECg3E9hTt20I7gh5GNZqAA7FVgBGoZw_O74JWY9coyrn9eis1hf94dU-JorqxNRD1ocg51vMqmkXlTFH9Fb1a2RRs</recordid><startdate>20230502</startdate><enddate>20230502</enddate><creator>Zhu, Min</creator><creator>Zhang, Handi</creator><creator>Jiao, Anran</creator><creator>Karniadakis, George Em</creator><creator>Lu, Lu</creator><general>Elsevier</general><scope>OIOZB</scope><scope>OTOTI</scope><orcidid>https://orcid.org/0000000254765768</orcidid><orcidid>https://orcid.org/0000000345505662</orcidid></search><sort><creationdate>20230502</creationdate><title>Reliable extrapolation of deep neural operators informed by physics or sparse observations</title><author>Zhu, Min ; Zhang, Handi ; Jiao, Anran ; Karniadakis, George Em ; Lu, Lu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-osti_scitechconnect_19912933</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>DeepONet</topic><topic>ENGINEERING</topic><topic>Extrapolation complexity</topic><topic>Fine-tuning</topic><topic>Multifidelity learning</topic><topic>Neural operators</topic><topic>Out-of-distribution inference</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhu, Min</creatorcontrib><creatorcontrib>Zhang, Handi</creatorcontrib><creatorcontrib>Jiao, Anran</creatorcontrib><creatorcontrib>Karniadakis, George Em</creatorcontrib><creatorcontrib>Lu, Lu</creatorcontrib><creatorcontrib>Univ. of Pennsylvania, Philadelphia, PA (United States)</creatorcontrib><collection>OSTI.GOV - Hybrid</collection><collection>OSTI.GOV</collection><jtitle>Computer methods in applied mechanics and engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhu, Min</au><au>Zhang, Handi</au><au>Jiao, Anran</au><au>Karniadakis, George Em</au><au>Lu, Lu</au><aucorp>Univ. of Pennsylvania, Philadelphia, PA (United States)</aucorp><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Reliable extrapolation of deep neural operators informed by physics or sparse observations</atitle><jtitle>Computer methods in applied mechanics and engineering</jtitle><date>2023-05-02</date><risdate>2023</risdate><volume>412</volume><issn>0045-7825</issn><eissn>1879-2138</eissn><abstract>Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks. As promising surrogate solvers of partial differential equations (PDEs) for real-time prediction, deep neural operators such as deep operator networks (DeepONets) provide a new simulation paradigm in science and engineering. Pure data-driven neural operators and deep learning models, in general, are usually limited to interpolation scenarios, where new predictions utilize inputs within the support of the training set. However, in the inference stage of real-world applications, the input may lie outside the support, i.e., extrapolation is required, which may result to large errors and unavoidable failure of deep learning models. Here, we address this challenge of extrapolation for deep neural operators. First, we systematically investigate the extrapolation behavior of DeepONets by quantifying the extrapolation complexity, via the 2-Wasserstein distance between two function spaces and propose a new strategy of bias–variance trade-off for extrapolation with respect to model capacity. Subsequently, we develop a complete workflow, including extrapolation determination, and we propose five reliable learning methods that guarantee a safe prediction under extrapolation by requiring additional information—the governing PDEs of the system or sparse new observations. The proposed methods are based on either fine-tuning a pre-trained DeepONet or multifidelity learning. We demonstrate the effectiveness of the proposed framework for various types of parametric PDEs. Furthermore, our systematic comparisons provide practical guidelines for selecting a proper extrapolation method depending on the available information, desired accuracy, and required inference speed.</abstract><cop>United States</cop><pub>Elsevier</pub><orcidid>https://orcid.org/0000000254765768</orcidid><orcidid>https://orcid.org/0000000345505662</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0045-7825
ispartof Computer methods in applied mechanics and engineering, 2023-05, Vol.412
issn 0045-7825
1879-2138
language eng
recordid cdi_osti_scitechconnect_1991293
source Elsevier ScienceDirect Journals
subjects DeepONet
ENGINEERING
Extrapolation complexity
Fine-tuning
Multifidelity learning
Neural operators
Out-of-distribution inference
title Reliable extrapolation of deep neural operators informed by physics or sparse observations
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T02%3A43%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-osti&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Reliable%20extrapolation%20of%20deep%20neural%20operators%20informed%20by%20physics%20or%20sparse%20observations&rft.jtitle=Computer%20methods%20in%20applied%20mechanics%20and%20engineering&rft.au=Zhu,%20Min&rft.aucorp=Univ.%20of%20Pennsylvania,%20Philadelphia,%20PA%20(United%20States)&rft.date=2023-05-02&rft.volume=412&rft.issn=0045-7825&rft.eissn=1879-2138&rft_id=info:doi/&rft_dat=%3Costi%3E1991293%3C/osti%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true