Model-Based Reinforcement Learning with Multinomial Logistic Function Approximation
We study model-based reinforcement learning (RL) for episodic Markov decision processes (MDP) whose transition probability is parametrized by an unknown transition core with features of state and action. Despite much recent progress in analyzing algorithms in the linear MDP setting, the understandin...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Hwang, Taehyun Oh, Min-hwan |
description | We study model-based reinforcement learning (RL) for episodic Markov decision
processes (MDP) whose transition probability is parametrized by an unknown
transition core with features of state and action. Despite much recent progress
in analyzing algorithms in the linear MDP setting, the understanding of more
general transition models is very restrictive. In this paper, we establish a
provably efficient RL algorithm for the MDP whose state transition is given by
a multinomial logistic model. To balance the exploration-exploitation
trade-off, we propose an upper confidence bound-based algorithm. We show that
our proposed algorithm achieves $\tilde{O}(d \sqrt{H^3 T})$ regret bound where
$d$ is the dimension of the transition core, $H$ is the horizon, and $T$ is the
total number of steps. To the best of our knowledge, this is the first
model-based RL algorithm with multinomial logistic function approximation with
provable guarantees. We also comprehensively evaluate our proposed algorithm
numerically and show that it consistently outperforms the existing methods,
hence achieving both provable efficiency and practical superior performance. |
doi_str_mv | 10.48550/arxiv.2212.13540 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2212_13540</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2212_13540</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-8658c87a844c721642f8e4ddda223afbc416665ef9f323dddc1cb3b30c29e54e3</originalsourceid><addsrcrecordid>eNotj01OwzAYRL1hgQoHYIUvkOD_uMtSUUBKhQTdR479uVhK7MhxodyetrAazSze6CF0R0kttJTkweRj-KoZo6ymXApyjT62ycFQPZoZHH6HEH3KFkaIBbdgcgxxj79D-cTbw1BCTGMwA27TPswlWLw5RFtCing1TTkdw2jO7QZdeTPMcPufC7TbPO3WL1X79vy6XrWVUQ2ptJLa6sZoIWzDqBLMaxDOOcMYN763giqlJPil54yfdkttz3tOLFuCFMAX6P4Pe7Hqpny6zz_d2a672PFfvRlLug</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Model-Based Reinforcement Learning with Multinomial Logistic Function Approximation</title><source>arXiv.org</source><creator>Hwang, Taehyun ; Oh, Min-hwan</creator><creatorcontrib>Hwang, Taehyun ; Oh, Min-hwan</creatorcontrib><description>We study model-based reinforcement learning (RL) for episodic Markov decision
processes (MDP) whose transition probability is parametrized by an unknown
transition core with features of state and action. Despite much recent progress
in analyzing algorithms in the linear MDP setting, the understanding of more
general transition models is very restrictive. In this paper, we establish a
provably efficient RL algorithm for the MDP whose state transition is given by
a multinomial logistic model. To balance the exploration-exploitation
trade-off, we propose an upper confidence bound-based algorithm. We show that
our proposed algorithm achieves $\tilde{O}(d \sqrt{H^3 T})$ regret bound where
$d$ is the dimension of the transition core, $H$ is the horizon, and $T$ is the
total number of steps. To the best of our knowledge, this is the first
model-based RL algorithm with multinomial logistic function approximation with
provable guarantees. We also comprehensively evaluate our proposed algorithm
numerically and show that it consistently outperforms the existing methods,
hence achieving both provable efficiency and practical superior performance.</description><identifier>DOI: 10.48550/arxiv.2212.13540</identifier><language>eng</language><subject>Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2022-12</creationdate><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2212.13540$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2212.13540$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Hwang, Taehyun</creatorcontrib><creatorcontrib>Oh, Min-hwan</creatorcontrib><title>Model-Based Reinforcement Learning with Multinomial Logistic Function Approximation</title><description>We study model-based reinforcement learning (RL) for episodic Markov decision
processes (MDP) whose transition probability is parametrized by an unknown
transition core with features of state and action. Despite much recent progress
in analyzing algorithms in the linear MDP setting, the understanding of more
general transition models is very restrictive. In this paper, we establish a
provably efficient RL algorithm for the MDP whose state transition is given by
a multinomial logistic model. To balance the exploration-exploitation
trade-off, we propose an upper confidence bound-based algorithm. We show that
our proposed algorithm achieves $\tilde{O}(d \sqrt{H^3 T})$ regret bound where
$d$ is the dimension of the transition core, $H$ is the horizon, and $T$ is the
total number of steps. To the best of our knowledge, this is the first
model-based RL algorithm with multinomial logistic function approximation with
provable guarantees. We also comprehensively evaluate our proposed algorithm
numerically and show that it consistently outperforms the existing methods,
hence achieving both provable efficiency and practical superior performance.</description><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj01OwzAYRL1hgQoHYIUvkOD_uMtSUUBKhQTdR479uVhK7MhxodyetrAazSze6CF0R0kttJTkweRj-KoZo6ymXApyjT62ycFQPZoZHH6HEH3KFkaIBbdgcgxxj79D-cTbw1BCTGMwA27TPswlWLw5RFtCing1TTkdw2jO7QZdeTPMcPufC7TbPO3WL1X79vy6XrWVUQ2ptJLa6sZoIWzDqBLMaxDOOcMYN763giqlJPil54yfdkttz3tOLFuCFMAX6P4Pe7Hqpny6zz_d2a672PFfvRlLug</recordid><startdate>20221227</startdate><enddate>20221227</enddate><creator>Hwang, Taehyun</creator><creator>Oh, Min-hwan</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20221227</creationdate><title>Model-Based Reinforcement Learning with Multinomial Logistic Function Approximation</title><author>Hwang, Taehyun ; Oh, Min-hwan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-8658c87a844c721642f8e4ddda223afbc416665ef9f323dddc1cb3b30c29e54e3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Hwang, Taehyun</creatorcontrib><creatorcontrib>Oh, Min-hwan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Hwang, Taehyun</au><au>Oh, Min-hwan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Model-Based Reinforcement Learning with Multinomial Logistic Function Approximation</atitle><date>2022-12-27</date><risdate>2022</risdate><abstract>We study model-based reinforcement learning (RL) for episodic Markov decision
processes (MDP) whose transition probability is parametrized by an unknown
transition core with features of state and action. Despite much recent progress
in analyzing algorithms in the linear MDP setting, the understanding of more
general transition models is very restrictive. In this paper, we establish a
provably efficient RL algorithm for the MDP whose state transition is given by
a multinomial logistic model. To balance the exploration-exploitation
trade-off, we propose an upper confidence bound-based algorithm. We show that
our proposed algorithm achieves $\tilde{O}(d \sqrt{H^3 T})$ regret bound where
$d$ is the dimension of the transition core, $H$ is the horizon, and $T$ is the
total number of steps. To the best of our knowledge, this is the first
model-based RL algorithm with multinomial logistic function approximation with
provable guarantees. We also comprehensively evaluate our proposed algorithm
numerically and show that it consistently outperforms the existing methods,
hence achieving both provable efficiency and practical superior performance.</abstract><doi>10.48550/arxiv.2212.13540</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2212.13540 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2212_13540 |
source | arXiv.org |
subjects | Computer Science - Learning Statistics - Machine Learning |
title | Model-Based Reinforcement Learning with Multinomial Logistic Function Approximation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-25T04%3A11%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Model-Based%20Reinforcement%20Learning%20with%20Multinomial%20Logistic%20Function%20Approximation&rft.au=Hwang,%20Taehyun&rft.date=2022-12-27&rft_id=info:doi/10.48550/arxiv.2212.13540&rft_dat=%3Carxiv_GOX%3E2212_13540%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |