First-Explore, then Exploit: Meta-Learning to Solve Hard Exploration-Exploitation Trade-Offs
Standard reinforcement learning (RL) agents never intelligently explore like a human (i.e. taking into account complex domain priors and adapting quickly based on previous exploration). Across episodes, RL agents struggle to perform even simple exploration strategies, for example systematic search t...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Norman, Ben Clune, Jeff |
description | Standard reinforcement learning (RL) agents never intelligently explore like
a human (i.e. taking into account complex domain priors and adapting quickly
based on previous exploration). Across episodes, RL agents struggle to perform
even simple exploration strategies, for example systematic search that avoids
exploring the same location multiple times. This poor exploration limits
performance on challenging domains. Meta-RL is a potential solution, as unlike
standard RL, meta-RL can learn to explore, and potentially learn highly complex
strategies far beyond those of standard RL, strategies such as experimenting in
early episodes to learn new skills, or conducting experiments to learn about
the current environment. Traditional meta-RL focuses on the problem of learning
to optimally balance exploration and exploitation to maximize the cumulative
reward of the episode sequence (e.g., aiming to maximize the total wins in a
tournament -- while also improving as a player). We identify a new challenge
with state-of-the-art cumulative-reward meta-RL methods. When optimal behavior
requires exploration that sacrifices immediate reward to enable higher
subsequent reward, existing state-of-the-art cumulative-reward meta-RL methods
become stuck on the local optimum of failing to explore. Our method,
First-Explore, overcomes this limitation by learning two policies: one to
solely explore, and one to solely exploit. When exploring requires forgoing
early-episode reward, First-Explore significantly outperforms existing
cumulative meta-RL methods. By identifying and solving the previously
unrecognized problem of forgoing reward in early episodes, First-Explore
represents a significant step towards developing meta-RL algorithms capable of
human-like exploration on a broader range of domains. |
doi_str_mv | 10.48550/arxiv.2307.02276 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2307_02276</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2307_02276</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2307_022763</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjYw1zMwMjI342SIccssKi7Rda0oyMkvStVRKMlIzVMA8zJLrBR8U0sSdX1SE4vyMvPSFUryFYLzc8pSFTwSi1IgiooSSzLz83ShGsAchZCixJRUXf-0tGIeBta0xJziVF4ozc0g7-Ya4uyhC3ZHfEFRZm5iUWU8yD3xYPcYE1YBAAirQLE</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>First-Explore, then Exploit: Meta-Learning to Solve Hard Exploration-Exploitation Trade-Offs</title><source>arXiv.org</source><creator>Norman, Ben ; Clune, Jeff</creator><creatorcontrib>Norman, Ben ; Clune, Jeff</creatorcontrib><description>Standard reinforcement learning (RL) agents never intelligently explore like
a human (i.e. taking into account complex domain priors and adapting quickly
based on previous exploration). Across episodes, RL agents struggle to perform
even simple exploration strategies, for example systematic search that avoids
exploring the same location multiple times. This poor exploration limits
performance on challenging domains. Meta-RL is a potential solution, as unlike
standard RL, meta-RL can learn to explore, and potentially learn highly complex
strategies far beyond those of standard RL, strategies such as experimenting in
early episodes to learn new skills, or conducting experiments to learn about
the current environment. Traditional meta-RL focuses on the problem of learning
to optimally balance exploration and exploitation to maximize the cumulative
reward of the episode sequence (e.g., aiming to maximize the total wins in a
tournament -- while also improving as a player). We identify a new challenge
with state-of-the-art cumulative-reward meta-RL methods. When optimal behavior
requires exploration that sacrifices immediate reward to enable higher
subsequent reward, existing state-of-the-art cumulative-reward meta-RL methods
become stuck on the local optimum of failing to explore. Our method,
First-Explore, overcomes this limitation by learning two policies: one to
solely explore, and one to solely exploit. When exploring requires forgoing
early-episode reward, First-Explore significantly outperforms existing
cumulative meta-RL methods. By identifying and solving the previously
unrecognized problem of forgoing reward in early episodes, First-Explore
represents a significant step towards developing meta-RL algorithms capable of
human-like exploration on a broader range of domains.</description><identifier>DOI: 10.48550/arxiv.2307.02276</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2023-07</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,782,887</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2307.02276$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2307.02276$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Norman, Ben</creatorcontrib><creatorcontrib>Clune, Jeff</creatorcontrib><title>First-Explore, then Exploit: Meta-Learning to Solve Hard Exploration-Exploitation Trade-Offs</title><description>Standard reinforcement learning (RL) agents never intelligently explore like
a human (i.e. taking into account complex domain priors and adapting quickly
based on previous exploration). Across episodes, RL agents struggle to perform
even simple exploration strategies, for example systematic search that avoids
exploring the same location multiple times. This poor exploration limits
performance on challenging domains. Meta-RL is a potential solution, as unlike
standard RL, meta-RL can learn to explore, and potentially learn highly complex
strategies far beyond those of standard RL, strategies such as experimenting in
early episodes to learn new skills, or conducting experiments to learn about
the current environment. Traditional meta-RL focuses on the problem of learning
to optimally balance exploration and exploitation to maximize the cumulative
reward of the episode sequence (e.g., aiming to maximize the total wins in a
tournament -- while also improving as a player). We identify a new challenge
with state-of-the-art cumulative-reward meta-RL methods. When optimal behavior
requires exploration that sacrifices immediate reward to enable higher
subsequent reward, existing state-of-the-art cumulative-reward meta-RL methods
become stuck on the local optimum of failing to explore. Our method,
First-Explore, overcomes this limitation by learning two policies: one to
solely explore, and one to solely exploit. When exploring requires forgoing
early-episode reward, First-Explore significantly outperforms existing
cumulative meta-RL methods. By identifying and solving the previously
unrecognized problem of forgoing reward in early episodes, First-Explore
represents a significant step towards developing meta-RL algorithms capable of
human-like exploration on a broader range of domains.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjYw1zMwMjI342SIccssKi7Rda0oyMkvStVRKMlIzVMA8zJLrBR8U0sSdX1SE4vyMvPSFUryFYLzc8pSFTwSi1IgiooSSzLz83ShGsAchZCixJRUXf-0tGIeBta0xJziVF4ozc0g7-Ya4uyhC3ZHfEFRZm5iUWU8yD3xYPcYE1YBAAirQLE</recordid><startdate>20230705</startdate><enddate>20230705</enddate><creator>Norman, Ben</creator><creator>Clune, Jeff</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230705</creationdate><title>First-Explore, then Exploit: Meta-Learning to Solve Hard Exploration-Exploitation Trade-Offs</title><author>Norman, Ben ; Clune, Jeff</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2307_022763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Norman, Ben</creatorcontrib><creatorcontrib>Clune, Jeff</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Norman, Ben</au><au>Clune, Jeff</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>First-Explore, then Exploit: Meta-Learning to Solve Hard Exploration-Exploitation Trade-Offs</atitle><date>2023-07-05</date><risdate>2023</risdate><abstract>Standard reinforcement learning (RL) agents never intelligently explore like
a human (i.e. taking into account complex domain priors and adapting quickly
based on previous exploration). Across episodes, RL agents struggle to perform
even simple exploration strategies, for example systematic search that avoids
exploring the same location multiple times. This poor exploration limits
performance on challenging domains. Meta-RL is a potential solution, as unlike
standard RL, meta-RL can learn to explore, and potentially learn highly complex
strategies far beyond those of standard RL, strategies such as experimenting in
early episodes to learn new skills, or conducting experiments to learn about
the current environment. Traditional meta-RL focuses on the problem of learning
to optimally balance exploration and exploitation to maximize the cumulative
reward of the episode sequence (e.g., aiming to maximize the total wins in a
tournament -- while also improving as a player). We identify a new challenge
with state-of-the-art cumulative-reward meta-RL methods. When optimal behavior
requires exploration that sacrifices immediate reward to enable higher
subsequent reward, existing state-of-the-art cumulative-reward meta-RL methods
become stuck on the local optimum of failing to explore. Our method,
First-Explore, overcomes this limitation by learning two policies: one to
solely explore, and one to solely exploit. When exploring requires forgoing
early-episode reward, First-Explore significantly outperforms existing
cumulative meta-RL methods. By identifying and solving the previously
unrecognized problem of forgoing reward in early episodes, First-Explore
represents a significant step towards developing meta-RL algorithms capable of
human-like exploration on a broader range of domains.</abstract><doi>10.48550/arxiv.2307.02276</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2307.02276 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2307_02276 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning |
title | First-Explore, then Exploit: Meta-Learning to Solve Hard Exploration-Exploitation Trade-Offs |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-02T11%3A02%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=First-Explore,%20then%20Exploit:%20Meta-Learning%20to%20Solve%20Hard%20Exploration-Exploitation%20Trade-Offs&rft.au=Norman,%20Ben&rft.date=2023-07-05&rft_id=info:doi/10.48550/arxiv.2307.02276&rft_dat=%3Carxiv_GOX%3E2307_02276%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |