Iterative construction of Gaussian process surrogate models for Bayesian inference

A new algorithm is developed to tackle the issue of sampling non-Gaussian model parameter posterior probability distributions that arise from solutions to Bayesian inverse problems. The algorithm aims to mitigate some of the hurdles faced by traditional Markov Chain Monte Carlo (MCMC) samplers, thro...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of statistical planning and inference 2020-07, Vol.207 (C), p.55-72
Hauptverfasser: Alawieh, Leen, Goodman, Jonathan, Bell, John B.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 72
container_issue C
container_start_page 55
container_title Journal of statistical planning and inference
container_volume 207
creator Alawieh, Leen
Goodman, Jonathan
Bell, John B.
description A new algorithm is developed to tackle the issue of sampling non-Gaussian model parameter posterior probability distributions that arise from solutions to Bayesian inverse problems. The algorithm aims to mitigate some of the hurdles faced by traditional Markov Chain Monte Carlo (MCMC) samplers, through constructing proposal probability densities that are both, easy to sample and that provide a better approximation to the target density than a simple Gaussian proposal distribution would. To achieve that, a Gaussian proposal distribution is augmented with a Gaussian Process (GP) surface that helps capture non-linearities in the log-likelihood function. In order to train the GP surface, an iterative approach is adopted for the optimal selection of points in parameter space. Optimality is sought by maximizing the information gain of the GP surface using a minimum number of forward model simulation runs. The accuracy of the GP-augmented surface approximation is assessed in two ways. The first consists of comparing predictions obtained from the approximate surface with those obtained through running the actual simulation model at hold-out points in parameter space. The second consists of a measure based on the relative variance of sample weights obtained from sampling the approximate posterior probability distribution of the model parameters. The efficacy of this new algorithm is tested on inferring reaction rate parameters in a 3-node and 6-node network toy problems, which imitate idealized reaction networks in combustion applications. •An adaptive emulator is built using Gaussian process regression.•An acquisition function is derived for optimal training of the emulator.•MCMC sampler is used to optimize the acquisition function.•Emulator is used to reduce computational cost of a Bayesian inverse problem.
doi_str_mv 10.1016/j.jspi.2019.11.002
format Article
fullrecord <record><control><sourceid>elsevier_osti_</sourceid><recordid>TN_cdi_osti_scitechconnect_1598372</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S037837581930103X</els_id><sourcerecordid>S037837581930103X</sourcerecordid><originalsourceid>FETCH-LOGICAL-c371t-de98cc7262e17bd0199583fc408060c11a43ffbbb60f7292ed02a45686542cb43</originalsourceid><addsrcrecordid>eNp9kFFLwzAUhYMoOKd_wKfie2tu0jYp-KJD52AgiD6HNL3RlK0ZSTbYv7d1Pntf7ss5h3M-Qm6BFkChvu-LPu5cwSg0BUBBKTsjM5CC5wACzsmMciFzLip5Sa5i7Ol4Na1m5H2VMOjkDpgZP8QU9iY5P2TeZku9j9HpIdsFbzDGLO5D8F86Ybb1HW5iZn3InvQRf1VusBhwMHhNLqzeRLz5-3Py-fL8sXjN12_L1eJxnRsuIOUdNtIYwWqGINpubN5UkltTUjlWMwC65Na2bVtTK1jDsKNMl1Ut66pkpi35nNydcn1MTkXjEprvccSAJimoGskFG0XsJDLBxxjQql1wWx2OCqia0KleTejUhE4BqBHdaHo4mcaReHAYpvRpWufCFN5595_9B1lgeGI</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Iterative construction of Gaussian process surrogate models for Bayesian inference</title><source>Elsevier ScienceDirect Journals Complete</source><creator>Alawieh, Leen ; Goodman, Jonathan ; Bell, John B.</creator><creatorcontrib>Alawieh, Leen ; Goodman, Jonathan ; Bell, John B.</creatorcontrib><description>A new algorithm is developed to tackle the issue of sampling non-Gaussian model parameter posterior probability distributions that arise from solutions to Bayesian inverse problems. The algorithm aims to mitigate some of the hurdles faced by traditional Markov Chain Monte Carlo (MCMC) samplers, through constructing proposal probability densities that are both, easy to sample and that provide a better approximation to the target density than a simple Gaussian proposal distribution would. To achieve that, a Gaussian proposal distribution is augmented with a Gaussian Process (GP) surface that helps capture non-linearities in the log-likelihood function. In order to train the GP surface, an iterative approach is adopted for the optimal selection of points in parameter space. Optimality is sought by maximizing the information gain of the GP surface using a minimum number of forward model simulation runs. The accuracy of the GP-augmented surface approximation is assessed in two ways. The first consists of comparing predictions obtained from the approximate surface with those obtained through running the actual simulation model at hold-out points in parameter space. The second consists of a measure based on the relative variance of sample weights obtained from sampling the approximate posterior probability distribution of the model parameters. The efficacy of this new algorithm is tested on inferring reaction rate parameters in a 3-node and 6-node network toy problems, which imitate idealized reaction networks in combustion applications. •An adaptive emulator is built using Gaussian process regression.•An acquisition function is derived for optimal training of the emulator.•MCMC sampler is used to optimize the acquisition function.•Emulator is used to reduce computational cost of a Bayesian inverse problem.</description><identifier>ISSN: 0378-3758</identifier><identifier>EISSN: 1873-1171</identifier><identifier>DOI: 10.1016/j.jspi.2019.11.002</identifier><language>eng</language><publisher>Elsevier B.V</publisher><subject>Active learning ; Bayesian inference ; Gaussian process regression ; MCMC ; Surrogate models</subject><ispartof>Journal of statistical planning and inference, 2020-07, Vol.207 (C), p.55-72</ispartof><rights>2019 Elsevier B.V.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c371t-de98cc7262e17bd0199583fc408060c11a43ffbbb60f7292ed02a45686542cb43</citedby><cites>FETCH-LOGICAL-c371t-de98cc7262e17bd0199583fc408060c11a43ffbbb60f7292ed02a45686542cb43</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.jspi.2019.11.002$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,776,780,881,3536,27903,27904,45974</link.rule.ids><backlink>$$Uhttps://www.osti.gov/biblio/1598372$$D View this record in Osti.gov$$Hfree_for_read</backlink></links><search><creatorcontrib>Alawieh, Leen</creatorcontrib><creatorcontrib>Goodman, Jonathan</creatorcontrib><creatorcontrib>Bell, John B.</creatorcontrib><title>Iterative construction of Gaussian process surrogate models for Bayesian inference</title><title>Journal of statistical planning and inference</title><description>A new algorithm is developed to tackle the issue of sampling non-Gaussian model parameter posterior probability distributions that arise from solutions to Bayesian inverse problems. The algorithm aims to mitigate some of the hurdles faced by traditional Markov Chain Monte Carlo (MCMC) samplers, through constructing proposal probability densities that are both, easy to sample and that provide a better approximation to the target density than a simple Gaussian proposal distribution would. To achieve that, a Gaussian proposal distribution is augmented with a Gaussian Process (GP) surface that helps capture non-linearities in the log-likelihood function. In order to train the GP surface, an iterative approach is adopted for the optimal selection of points in parameter space. Optimality is sought by maximizing the information gain of the GP surface using a minimum number of forward model simulation runs. The accuracy of the GP-augmented surface approximation is assessed in two ways. The first consists of comparing predictions obtained from the approximate surface with those obtained through running the actual simulation model at hold-out points in parameter space. The second consists of a measure based on the relative variance of sample weights obtained from sampling the approximate posterior probability distribution of the model parameters. The efficacy of this new algorithm is tested on inferring reaction rate parameters in a 3-node and 6-node network toy problems, which imitate idealized reaction networks in combustion applications. •An adaptive emulator is built using Gaussian process regression.•An acquisition function is derived for optimal training of the emulator.•MCMC sampler is used to optimize the acquisition function.•Emulator is used to reduce computational cost of a Bayesian inverse problem.</description><subject>Active learning</subject><subject>Bayesian inference</subject><subject>Gaussian process regression</subject><subject>MCMC</subject><subject>Surrogate models</subject><issn>0378-3758</issn><issn>1873-1171</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNp9kFFLwzAUhYMoOKd_wKfie2tu0jYp-KJD52AgiD6HNL3RlK0ZSTbYv7d1Pntf7ss5h3M-Qm6BFkChvu-LPu5cwSg0BUBBKTsjM5CC5wACzsmMciFzLip5Sa5i7Ol4Na1m5H2VMOjkDpgZP8QU9iY5P2TeZku9j9HpIdsFbzDGLO5D8F86Ybb1HW5iZn3InvQRf1VusBhwMHhNLqzeRLz5-3Py-fL8sXjN12_L1eJxnRsuIOUdNtIYwWqGINpubN5UkltTUjlWMwC65Na2bVtTK1jDsKNMl1Ut66pkpi35nNydcn1MTkXjEprvccSAJimoGskFG0XsJDLBxxjQql1wWx2OCqia0KleTejUhE4BqBHdaHo4mcaReHAYpvRpWufCFN5595_9B1lgeGI</recordid><startdate>202007</startdate><enddate>202007</enddate><creator>Alawieh, Leen</creator><creator>Goodman, Jonathan</creator><creator>Bell, John B.</creator><general>Elsevier B.V</general><general>Elsevier</general><scope>AAYXX</scope><scope>CITATION</scope><scope>OTOTI</scope></search><sort><creationdate>202007</creationdate><title>Iterative construction of Gaussian process surrogate models for Bayesian inference</title><author>Alawieh, Leen ; Goodman, Jonathan ; Bell, John B.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c371t-de98cc7262e17bd0199583fc408060c11a43ffbbb60f7292ed02a45686542cb43</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Active learning</topic><topic>Bayesian inference</topic><topic>Gaussian process regression</topic><topic>MCMC</topic><topic>Surrogate models</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Alawieh, Leen</creatorcontrib><creatorcontrib>Goodman, Jonathan</creatorcontrib><creatorcontrib>Bell, John B.</creatorcontrib><collection>CrossRef</collection><collection>OSTI.GOV</collection><jtitle>Journal of statistical planning and inference</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Alawieh, Leen</au><au>Goodman, Jonathan</au><au>Bell, John B.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Iterative construction of Gaussian process surrogate models for Bayesian inference</atitle><jtitle>Journal of statistical planning and inference</jtitle><date>2020-07</date><risdate>2020</risdate><volume>207</volume><issue>C</issue><spage>55</spage><epage>72</epage><pages>55-72</pages><issn>0378-3758</issn><eissn>1873-1171</eissn><abstract>A new algorithm is developed to tackle the issue of sampling non-Gaussian model parameter posterior probability distributions that arise from solutions to Bayesian inverse problems. The algorithm aims to mitigate some of the hurdles faced by traditional Markov Chain Monte Carlo (MCMC) samplers, through constructing proposal probability densities that are both, easy to sample and that provide a better approximation to the target density than a simple Gaussian proposal distribution would. To achieve that, a Gaussian proposal distribution is augmented with a Gaussian Process (GP) surface that helps capture non-linearities in the log-likelihood function. In order to train the GP surface, an iterative approach is adopted for the optimal selection of points in parameter space. Optimality is sought by maximizing the information gain of the GP surface using a minimum number of forward model simulation runs. The accuracy of the GP-augmented surface approximation is assessed in two ways. The first consists of comparing predictions obtained from the approximate surface with those obtained through running the actual simulation model at hold-out points in parameter space. The second consists of a measure based on the relative variance of sample weights obtained from sampling the approximate posterior probability distribution of the model parameters. The efficacy of this new algorithm is tested on inferring reaction rate parameters in a 3-node and 6-node network toy problems, which imitate idealized reaction networks in combustion applications. •An adaptive emulator is built using Gaussian process regression.•An acquisition function is derived for optimal training of the emulator.•MCMC sampler is used to optimize the acquisition function.•Emulator is used to reduce computational cost of a Bayesian inverse problem.</abstract><pub>Elsevier B.V</pub><doi>10.1016/j.jspi.2019.11.002</doi><tpages>18</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0378-3758
ispartof Journal of statistical planning and inference, 2020-07, Vol.207 (C), p.55-72
issn 0378-3758
1873-1171
language eng
recordid cdi_osti_scitechconnect_1598372
source Elsevier ScienceDirect Journals Complete
subjects Active learning
Bayesian inference
Gaussian process regression
MCMC
Surrogate models
title Iterative construction of Gaussian process surrogate models for Bayesian inference
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T09%3A10%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_osti_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Iterative%20construction%20of%20Gaussian%20process%20surrogate%20models%20for%20Bayesian%20inference&rft.jtitle=Journal%20of%20statistical%20planning%20and%20inference&rft.au=Alawieh,%20Leen&rft.date=2020-07&rft.volume=207&rft.issue=C&rft.spage=55&rft.epage=72&rft.pages=55-72&rft.issn=0378-3758&rft.eissn=1873-1171&rft_id=info:doi/10.1016/j.jspi.2019.11.002&rft_dat=%3Celsevier_osti_%3ES037837581930103X%3C/elsevier_osti_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_els_id=S037837581930103X&rfr_iscdi=true