Deep Reinforcement Learning With Macro-Actions
Deep reinforcement learning has been shown to be a powerful framework for learning policies from complex high-dimensional sensory inputs to actions in complex tasks, such as the Atari domain. In this paper, we explore output representation modeling in the form of temporal abstraction to improve conv...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Durugkar, Ishan P Rosenbaum, Clemens Dernbach, Stefan Mahadevan, Sridhar |
description | Deep reinforcement learning has been shown to be a powerful framework for
learning policies from complex high-dimensional sensory inputs to actions in
complex tasks, such as the Atari domain. In this paper, we explore output
representation modeling in the form of temporal abstraction to improve
convergence and reliability of deep reinforcement learning approaches. We
concentrate on macro-actions, and evaluate these on different Atari 2600 games,
where we show that they yield significant improvements in learning speed.
Additionally, we show that they can even achieve better scores than DQN. We
offer analysis and explanation for both convergence and final results,
revealing a problem deep RL approaches have with sparse reward signals. |
doi_str_mv | 10.48550/arxiv.1606.04615 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1606_04615</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1606_04615</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-95859ca4c4d7cd3332f2b989b801e89ac43e2b23d31fce09269d2bb764c627af3</originalsourceid><addsrcrecordid>eNotzstqwkAUgOHZdFFsH6Ar8wJJ556ZpdiLQqRQBJfhzMkZHagTGUNp375UXf27n4-xJ8Eb7Yzhz1B-0ncjLLcN11aYe9a8EJ2qT0o5jgXpSHmqOoKSU95XuzQdqg1gGesFTmnM5wd2F-HrTI-3ztj27XW7XNXdx_t6uehqsK2pvXHGI2jUQ4uDUkpGGbzzwXFBzgNqRTJINSgRkbiX1g8yhNZqtLKFqGZsft1ewP2ppCOU3_4f3l_g6g9gADzL</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Deep Reinforcement Learning With Macro-Actions</title><source>arXiv.org</source><creator>Durugkar, Ishan P ; Rosenbaum, Clemens ; Dernbach, Stefan ; Mahadevan, Sridhar</creator><creatorcontrib>Durugkar, Ishan P ; Rosenbaum, Clemens ; Dernbach, Stefan ; Mahadevan, Sridhar</creatorcontrib><description>Deep reinforcement learning has been shown to be a powerful framework for
learning policies from complex high-dimensional sensory inputs to actions in
complex tasks, such as the Atari domain. In this paper, we explore output
representation modeling in the form of temporal abstraction to improve
convergence and reliability of deep reinforcement learning approaches. We
concentrate on macro-actions, and evaluate these on different Atari 2600 games,
where we show that they yield significant improvements in learning speed.
Additionally, we show that they can even achieve better scores than DQN. We
offer analysis and explanation for both convergence and final results,
revealing a problem deep RL approaches have with sparse reward signals.</description><identifier>DOI: 10.48550/arxiv.1606.04615</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Computer Science - Neural and Evolutionary Computing</subject><creationdate>2016-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,778,883</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1606.04615$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1606.04615$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Durugkar, Ishan P</creatorcontrib><creatorcontrib>Rosenbaum, Clemens</creatorcontrib><creatorcontrib>Dernbach, Stefan</creatorcontrib><creatorcontrib>Mahadevan, Sridhar</creatorcontrib><title>Deep Reinforcement Learning With Macro-Actions</title><description>Deep reinforcement learning has been shown to be a powerful framework for
learning policies from complex high-dimensional sensory inputs to actions in
complex tasks, such as the Atari domain. In this paper, we explore output
representation modeling in the form of temporal abstraction to improve
convergence and reliability of deep reinforcement learning approaches. We
concentrate on macro-actions, and evaluate these on different Atari 2600 games,
where we show that they yield significant improvements in learning speed.
Additionally, we show that they can even achieve better scores than DQN. We
offer analysis and explanation for both convergence and final results,
revealing a problem deep RL approaches have with sparse reward signals.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Neural and Evolutionary Computing</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2016</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzstqwkAUgOHZdFFsH6Ar8wJJ556ZpdiLQqRQBJfhzMkZHagTGUNp375UXf27n4-xJ8Eb7Yzhz1B-0ncjLLcN11aYe9a8EJ2qT0o5jgXpSHmqOoKSU95XuzQdqg1gGesFTmnM5wd2F-HrTI-3ztj27XW7XNXdx_t6uehqsK2pvXHGI2jUQ4uDUkpGGbzzwXFBzgNqRTJINSgRkbiX1g8yhNZqtLKFqGZsft1ewP2ppCOU3_4f3l_g6g9gADzL</recordid><startdate>20160614</startdate><enddate>20160614</enddate><creator>Durugkar, Ishan P</creator><creator>Rosenbaum, Clemens</creator><creator>Dernbach, Stefan</creator><creator>Mahadevan, Sridhar</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20160614</creationdate><title>Deep Reinforcement Learning With Macro-Actions</title><author>Durugkar, Ishan P ; Rosenbaum, Clemens ; Dernbach, Stefan ; Mahadevan, Sridhar</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-95859ca4c4d7cd3332f2b989b801e89ac43e2b23d31fce09269d2bb764c627af3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2016</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Neural and Evolutionary Computing</topic><toplevel>online_resources</toplevel><creatorcontrib>Durugkar, Ishan P</creatorcontrib><creatorcontrib>Rosenbaum, Clemens</creatorcontrib><creatorcontrib>Dernbach, Stefan</creatorcontrib><creatorcontrib>Mahadevan, Sridhar</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Durugkar, Ishan P</au><au>Rosenbaum, Clemens</au><au>Dernbach, Stefan</au><au>Mahadevan, Sridhar</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Reinforcement Learning With Macro-Actions</atitle><date>2016-06-14</date><risdate>2016</risdate><abstract>Deep reinforcement learning has been shown to be a powerful framework for
learning policies from complex high-dimensional sensory inputs to actions in
complex tasks, such as the Atari domain. In this paper, we explore output
representation modeling in the form of temporal abstraction to improve
convergence and reliability of deep reinforcement learning approaches. We
concentrate on macro-actions, and evaluate these on different Atari 2600 games,
where we show that they yield significant improvements in learning speed.
Additionally, we show that they can even achieve better scores than DQN. We
offer analysis and explanation for both convergence and final results,
revealing a problem deep RL approaches have with sparse reward signals.</abstract><doi>10.48550/arxiv.1606.04615</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1606.04615 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1606_04615 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning Computer Science - Neural and Evolutionary Computing |
title | Deep Reinforcement Learning With Macro-Actions |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T09%3A28%3A15IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Reinforcement%20Learning%20With%20Macro-Actions&rft.au=Durugkar,%20Ishan%20P&rft.date=2016-06-14&rft_id=info:doi/10.48550/arxiv.1606.04615&rft_dat=%3Carxiv_GOX%3E1606_04615%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |