Newton's Method and Differential Dynamic Programming for Unconstrained Nonlinear Dynamic Games
Dynamic games arise when multiple agents with differing objectives control a dynamic system. They model a wide variety of applications in economics, defense, energy systems and etc. However, compared to single-agent control problems, the computational methods for dynamic games are relatively limited...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2020-01 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Bolei Di Lamperski, Andrew |
description | Dynamic games arise when multiple agents with differing objectives control a dynamic system. They model a wide variety of applications in economics, defense, energy systems and etc. However, compared to single-agent control problems, the computational methods for dynamic games are relatively limited. As in the single-agent case, only specific dynamic games can be solved exactly, so approximation algorithms are required. In this paper, we show how to extend a recursive Newton's algorithm and the popular differential dynamic programming (DDP) for single-agent optimal control to the case of full-information non-zero sum dynamic games. In the single-agent case, the convergence of DDP is proved by comparison with Newton's method, which converges locally at a quadratic rate. We show that the iterates of Newton's method and DDP are sufficiently close for the DDP to inherit the quadratic convergence rate of Newton's method. We also prove both methods result in an open-loop Nash equilibrium and a local feedback \(O(\epsilon^2)\)-Nash equilibrium. Numerical examples are provided. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2245977862</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2245977862</sourcerecordid><originalsourceid>FETCH-proquest_journals_22459778623</originalsourceid><addsrcrecordid>eNqNirsKwjAUQIMgWLT_EHBwKtSkL2frY7E46GoJbVpTmhu9SRH_3g7i7HQOnDMhHuN8HWQRYzPiW9uFYciSlMUx98itkC9nYGXpSbq7qamAmuaqaSRKcEr0NH-D0KqiZzQtCq0VtLQxSK9QGbAOhQJZ08JAP4rA338QWtoFmTait9L_ck6W-91lewweaJ6DtK7szIAwppKxKN6kaZYw_t_1AVXQREU</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2245977862</pqid></control><display><type>article</type><title>Newton's Method and Differential Dynamic Programming for Unconstrained Nonlinear Dynamic Games</title><source>Free E- Journals</source><creator>Bolei Di ; Lamperski, Andrew</creator><creatorcontrib>Bolei Di ; Lamperski, Andrew</creatorcontrib><description>Dynamic games arise when multiple agents with differing objectives control a dynamic system. They model a wide variety of applications in economics, defense, energy systems and etc. However, compared to single-agent control problems, the computational methods for dynamic games are relatively limited. As in the single-agent case, only specific dynamic games can be solved exactly, so approximation algorithms are required. In this paper, we show how to extend a recursive Newton's algorithm and the popular differential dynamic programming (DDP) for single-agent optimal control to the case of full-information non-zero sum dynamic games. In the single-agent case, the convergence of DDP is proved by comparison with Newton's method, which converges locally at a quadratic rate. We show that the iterates of Newton's method and DDP are sufficiently close for the DDP to inherit the quadratic convergence rate of Newton's method. We also prove both methods result in an open-loop Nash equilibrium and a local feedback \(O(\epsilon^2)\)-Nash equilibrium. Numerical examples are provided.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Control methods ; Convergence ; Dynamic programming ; Dynamical systems ; Economic models ; Games ; Multiagent systems ; Newton methods ; Nonlinear dynamics ; Optimal control</subject><ispartof>arXiv.org, 2020-01</ispartof><rights>2020. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Bolei Di</creatorcontrib><creatorcontrib>Lamperski, Andrew</creatorcontrib><title>Newton's Method and Differential Dynamic Programming for Unconstrained Nonlinear Dynamic Games</title><title>arXiv.org</title><description>Dynamic games arise when multiple agents with differing objectives control a dynamic system. They model a wide variety of applications in economics, defense, energy systems and etc. However, compared to single-agent control problems, the computational methods for dynamic games are relatively limited. As in the single-agent case, only specific dynamic games can be solved exactly, so approximation algorithms are required. In this paper, we show how to extend a recursive Newton's algorithm and the popular differential dynamic programming (DDP) for single-agent optimal control to the case of full-information non-zero sum dynamic games. In the single-agent case, the convergence of DDP is proved by comparison with Newton's method, which converges locally at a quadratic rate. We show that the iterates of Newton's method and DDP are sufficiently close for the DDP to inherit the quadratic convergence rate of Newton's method. We also prove both methods result in an open-loop Nash equilibrium and a local feedback \(O(\epsilon^2)\)-Nash equilibrium. Numerical examples are provided.</description><subject>Algorithms</subject><subject>Control methods</subject><subject>Convergence</subject><subject>Dynamic programming</subject><subject>Dynamical systems</subject><subject>Economic models</subject><subject>Games</subject><subject>Multiagent systems</subject><subject>Newton methods</subject><subject>Nonlinear dynamics</subject><subject>Optimal control</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNirsKwjAUQIMgWLT_EHBwKtSkL2frY7E46GoJbVpTmhu9SRH_3g7i7HQOnDMhHuN8HWQRYzPiW9uFYciSlMUx98itkC9nYGXpSbq7qamAmuaqaSRKcEr0NH-D0KqiZzQtCq0VtLQxSK9QGbAOhQJZ08JAP4rA338QWtoFmTait9L_ck6W-91lewweaJ6DtK7szIAwppKxKN6kaZYw_t_1AVXQREU</recordid><startdate>20200106</startdate><enddate>20200106</enddate><creator>Bolei Di</creator><creator>Lamperski, Andrew</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20200106</creationdate><title>Newton's Method and Differential Dynamic Programming for Unconstrained Nonlinear Dynamic Games</title><author>Bolei Di ; Lamperski, Andrew</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_22459778623</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Algorithms</topic><topic>Control methods</topic><topic>Convergence</topic><topic>Dynamic programming</topic><topic>Dynamical systems</topic><topic>Economic models</topic><topic>Games</topic><topic>Multiagent systems</topic><topic>Newton methods</topic><topic>Nonlinear dynamics</topic><topic>Optimal control</topic><toplevel>online_resources</toplevel><creatorcontrib>Bolei Di</creatorcontrib><creatorcontrib>Lamperski, Andrew</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Bolei Di</au><au>Lamperski, Andrew</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Newton's Method and Differential Dynamic Programming for Unconstrained Nonlinear Dynamic Games</atitle><jtitle>arXiv.org</jtitle><date>2020-01-06</date><risdate>2020</risdate><eissn>2331-8422</eissn><abstract>Dynamic games arise when multiple agents with differing objectives control a dynamic system. They model a wide variety of applications in economics, defense, energy systems and etc. However, compared to single-agent control problems, the computational methods for dynamic games are relatively limited. As in the single-agent case, only specific dynamic games can be solved exactly, so approximation algorithms are required. In this paper, we show how to extend a recursive Newton's algorithm and the popular differential dynamic programming (DDP) for single-agent optimal control to the case of full-information non-zero sum dynamic games. In the single-agent case, the convergence of DDP is proved by comparison with Newton's method, which converges locally at a quadratic rate. We show that the iterates of Newton's method and DDP are sufficiently close for the DDP to inherit the quadratic convergence rate of Newton's method. We also prove both methods result in an open-loop Nash equilibrium and a local feedback \(O(\epsilon^2)\)-Nash equilibrium. Numerical examples are provided.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2020-01 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2245977862 |
source | Free E- Journals |
subjects | Algorithms Control methods Convergence Dynamic programming Dynamical systems Economic models Games Multiagent systems Newton methods Nonlinear dynamics Optimal control |
title | Newton's Method and Differential Dynamic Programming for Unconstrained Nonlinear Dynamic Games |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T08%3A26%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Newton's%20Method%20and%20Differential%20Dynamic%20Programming%20for%20Unconstrained%20Nonlinear%20Dynamic%20Games&rft.jtitle=arXiv.org&rft.au=Bolei%20Di&rft.date=2020-01-06&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2245977862%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2245977862&rft_id=info:pmid/&rfr_iscdi=true |