Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space
Optimal control of diffusion processes is intimately connected to the problem of solving certain Hamilton-Jacobi-Bellman equations. Building on recent machine learning inspired approaches towards high-dimensional PDEs, we investigate the potential of $\textit{iterative diffusion optimisation}$ techn...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Nüsken, Nikolas Richter, Lorenz |
description | Optimal control of diffusion processes is intimately connected to the problem
of solving certain Hamilton-Jacobi-Bellman equations. Building on recent
machine learning inspired approaches towards high-dimensional PDEs, we
investigate the potential of $\textit{iterative diffusion optimisation}$
techniques, in particular considering applications in importance sampling and
rare event simulation, and focusing on problems without diffusion control, with
linearly controlled drift and running costs that depend quadratically on the
control. More generally, our methods apply to nonlinear parabolic PDEs with a
certain shift invariance. The choice of an appropriate loss function being a
central element in the algorithmic design, we develop a principled framework
based on divergences between path measures, encompassing various existing
methods. Motivated by connections to forward-backward SDEs, we propose and
study the novel $\textit{log-variance}$ divergence, showing favourable
properties of corresponding Monte Carlo estimators. The promise of the
developed approach is exemplified by a range of high-dimensional and metastable
numerical examples. |
doi_str_mv | 10.48550/arxiv.2005.05409 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2005_05409</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2005_05409</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-eecd1260c9f63e2a36d29ef397aaebaff5dbedfe443523a6047cf5c7dbb8193a3</originalsourceid><addsrcrecordid>eNotkE1OwzAQhbNhgQoHYMVcIMWN46RhB6XQokog0X01sceNhWNHdlLoYbgrTWHx9DbvR_qS5GbGpvlcCHaH4dscphljYspEzqrL5OfD24Nxe2jMvkmVaclF4x1aWGFrbO9d-orS1yZ9JGtbdPD-tIwwxLHjaAinpKP-y4fPeA8dhdiR7M2BIujgW-gbGuXDEbwG6V0fvLWkQBmth_EqAjoFLWEcwqnlHXTYNxA7lHSVXGi0ka7_fZJsn5fbxSrdvL2sFw-bFIuySomkmmUFk5UuOGXIC5VVpHlVIlKNWgtVk9KU51xkHAuWl1ILWaq6ns8qjnyS3P7NnvnsumBaDMfdyGl35sR_ARQ7aIk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space</title><source>arXiv.org</source><creator>Nüsken, Nikolas ; Richter, Lorenz</creator><creatorcontrib>Nüsken, Nikolas ; Richter, Lorenz</creatorcontrib><description>Optimal control of diffusion processes is intimately connected to the problem
of solving certain Hamilton-Jacobi-Bellman equations. Building on recent
machine learning inspired approaches towards high-dimensional PDEs, we
investigate the potential of $\textit{iterative diffusion optimisation}$
techniques, in particular considering applications in importance sampling and
rare event simulation, and focusing on problems without diffusion control, with
linearly controlled drift and running costs that depend quadratically on the
control. More generally, our methods apply to nonlinear parabolic PDEs with a
certain shift invariance. The choice of an appropriate loss function being a
central element in the algorithmic design, we develop a principled framework
based on divergences between path measures, encompassing various existing
methods. Motivated by connections to forward-backward SDEs, we propose and
study the novel $\textit{log-variance}$ divergence, showing favourable
properties of corresponding Monte Carlo estimators. The promise of the
developed approach is exemplified by a range of high-dimensional and metastable
numerical examples.</description><identifier>DOI: 10.48550/arxiv.2005.05409</identifier><language>eng</language><subject>Computer Science - Learning ; Computer Science - Numerical Analysis ; Mathematics - Numerical Analysis ; Mathematics - Optimization and Control ; Mathematics - Probability ; Statistics - Machine Learning</subject><creationdate>2020-05</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,882</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2005.05409$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2005.05409$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Nüsken, Nikolas</creatorcontrib><creatorcontrib>Richter, Lorenz</creatorcontrib><title>Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space</title><description>Optimal control of diffusion processes is intimately connected to the problem
of solving certain Hamilton-Jacobi-Bellman equations. Building on recent
machine learning inspired approaches towards high-dimensional PDEs, we
investigate the potential of $\textit{iterative diffusion optimisation}$
techniques, in particular considering applications in importance sampling and
rare event simulation, and focusing on problems without diffusion control, with
linearly controlled drift and running costs that depend quadratically on the
control. More generally, our methods apply to nonlinear parabolic PDEs with a
certain shift invariance. The choice of an appropriate loss function being a
central element in the algorithmic design, we develop a principled framework
based on divergences between path measures, encompassing various existing
methods. Motivated by connections to forward-backward SDEs, we propose and
study the novel $\textit{log-variance}$ divergence, showing favourable
properties of corresponding Monte Carlo estimators. The promise of the
developed approach is exemplified by a range of high-dimensional and metastable
numerical examples.</description><subject>Computer Science - Learning</subject><subject>Computer Science - Numerical Analysis</subject><subject>Mathematics - Numerical Analysis</subject><subject>Mathematics - Optimization and Control</subject><subject>Mathematics - Probability</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotkE1OwzAQhbNhgQoHYMVcIMWN46RhB6XQokog0X01sceNhWNHdlLoYbgrTWHx9DbvR_qS5GbGpvlcCHaH4dscphljYspEzqrL5OfD24Nxe2jMvkmVaclF4x1aWGFrbO9d-orS1yZ9JGtbdPD-tIwwxLHjaAinpKP-y4fPeA8dhdiR7M2BIujgW-gbGuXDEbwG6V0fvLWkQBmth_EqAjoFLWEcwqnlHXTYNxA7lHSVXGi0ka7_fZJsn5fbxSrdvL2sFw-bFIuySomkmmUFk5UuOGXIC5VVpHlVIlKNWgtVk9KU51xkHAuWl1ILWaq6ns8qjnyS3P7NnvnsumBaDMfdyGl35sR_ARQ7aIk</recordid><startdate>20200511</startdate><enddate>20200511</enddate><creator>Nüsken, Nikolas</creator><creator>Richter, Lorenz</creator><scope>AKY</scope><scope>AKZ</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20200511</creationdate><title>Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space</title><author>Nüsken, Nikolas ; Richter, Lorenz</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-eecd1260c9f63e2a36d29ef397aaebaff5dbedfe443523a6047cf5c7dbb8193a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Learning</topic><topic>Computer Science - Numerical Analysis</topic><topic>Mathematics - Numerical Analysis</topic><topic>Mathematics - Optimization and Control</topic><topic>Mathematics - Probability</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Nüsken, Nikolas</creatorcontrib><creatorcontrib>Richter, Lorenz</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Mathematics</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Nüsken, Nikolas</au><au>Richter, Lorenz</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space</atitle><date>2020-05-11</date><risdate>2020</risdate><abstract>Optimal control of diffusion processes is intimately connected to the problem
of solving certain Hamilton-Jacobi-Bellman equations. Building on recent
machine learning inspired approaches towards high-dimensional PDEs, we
investigate the potential of $\textit{iterative diffusion optimisation}$
techniques, in particular considering applications in importance sampling and
rare event simulation, and focusing on problems without diffusion control, with
linearly controlled drift and running costs that depend quadratically on the
control. More generally, our methods apply to nonlinear parabolic PDEs with a
certain shift invariance. The choice of an appropriate loss function being a
central element in the algorithmic design, we develop a principled framework
based on divergences between path measures, encompassing various existing
methods. Motivated by connections to forward-backward SDEs, we propose and
study the novel $\textit{log-variance}$ divergence, showing favourable
properties of corresponding Monte Carlo estimators. The promise of the
developed approach is exemplified by a range of high-dimensional and metastable
numerical examples.</abstract><doi>10.48550/arxiv.2005.05409</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2005.05409 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2005_05409 |
source | arXiv.org |
subjects | Computer Science - Learning Computer Science - Numerical Analysis Mathematics - Numerical Analysis Mathematics - Optimization and Control Mathematics - Probability Statistics - Machine Learning |
title | Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-17T22%3A19%3A30IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Solving%20high-dimensional%20Hamilton-Jacobi-Bellman%20PDEs%20using%20neural%20networks:%20perspectives%20from%20the%20theory%20of%20controlled%20diffusions%20and%20measures%20on%20path%20space&rft.au=N%C3%BCsken,%20Nikolas&rft.date=2020-05-11&rft_id=info:doi/10.48550/arxiv.2005.05409&rft_dat=%3Carxiv_GOX%3E2005_05409%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |