Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation
In recent years, domains such as natural language processing and image recognition have popularized the paradigm of using large datasets to pretrain representations that can be effectively transferred to downstream tasks. In this work we evaluate how such a paradigm should be done in imitation learn...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Brandfonbrener, David Nachum, Ofir Bruna, Joan |
description | In recent years, domains such as natural language processing and image
recognition have popularized the paradigm of using large datasets to pretrain
representations that can be effectively transferred to downstream tasks. In
this work we evaluate how such a paradigm should be done in imitation learning,
where both pretraining and finetuning data are trajectories collected by
experts interacting with an unknown environment. Namely, we consider a setting
where the pretraining corpus consists of multitask demonstrations and the task
for each demonstration is set by an unobserved latent context variable. The
goal is to use the pretraining corpus to learn a low dimensional representation
of the high dimensional (e.g., visual) observation space which can be
transferred to a novel context for finetuning on a limited dataset of
demonstrations. Among a variety of possible pretraining objectives, we argue
that inverse dynamics modeling -- i.e., predicting an action given the
observations appearing before and after it in the demonstration -- is
well-suited to this setting. We provide empirical evidence of this claim
through evaluations on a variety of simulated visuomotor manipulation problems.
While previous work has attempted various theoretical explanations regarding
the benefit of inverse dynamics modeling, we find that these arguments are
insufficient to explain the empirical advantages often observed in our
settings, and so we derive a novel analysis using a simple but general
environment model. |
doi_str_mv | 10.48550/arxiv.2305.16985 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2305_16985</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2305_16985</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-de767dea1e7512e424390cc252147d265571709d41dc1b1957c57dfe402402a93</originalsourceid><addsrcrecordid>eNotj1FLwzAUhfPig0x_gE_mD7QmaW6zPsrUWehwyN7LNbkdYWs6kjjcv3duwoED54MDH2MPUpR6DiCeMP74Y6kqAaWsmzncsnUbjhQT8ZdTwNHbxNeRckQffNjyjjCGxJfT5PgnHSIlChmzn87jMEW--t5nnzHteDv6K7hjNwPuE93_94xt3l43i_ei-1i2i-euwNpA4cjUxhFKMiAVaaWrRlirQEltnKoBjDSicVo6K79kA8aCcQNpoc7Bppqxx-vtRak_RD9iPPV_av1FrfoFlfZJbw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation</title><source>arXiv.org</source><creator>Brandfonbrener, David ; Nachum, Ofir ; Bruna, Joan</creator><creatorcontrib>Brandfonbrener, David ; Nachum, Ofir ; Bruna, Joan</creatorcontrib><description>In recent years, domains such as natural language processing and image
recognition have popularized the paradigm of using large datasets to pretrain
representations that can be effectively transferred to downstream tasks. In
this work we evaluate how such a paradigm should be done in imitation learning,
where both pretraining and finetuning data are trajectories collected by
experts interacting with an unknown environment. Namely, we consider a setting
where the pretraining corpus consists of multitask demonstrations and the task
for each demonstration is set by an unobserved latent context variable. The
goal is to use the pretraining corpus to learn a low dimensional representation
of the high dimensional (e.g., visual) observation space which can be
transferred to a novel context for finetuning on a limited dataset of
demonstrations. Among a variety of possible pretraining objectives, we argue
that inverse dynamics modeling -- i.e., predicting an action given the
observations appearing before and after it in the demonstration -- is
well-suited to this setting. We provide empirical evidence of this claim
through evaluations on a variety of simulated visuomotor manipulation problems.
While previous work has attempted various theoretical explanations regarding
the benefit of inverse dynamics modeling, we find that these arguments are
insufficient to explain the empirical advantages often observed in our
settings, and so we derive a novel analysis using a simple but general
environment model.</description><identifier>DOI: 10.48550/arxiv.2305.16985</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2023-05</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2305.16985$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2305.16985$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Brandfonbrener, David</creatorcontrib><creatorcontrib>Nachum, Ofir</creatorcontrib><creatorcontrib>Bruna, Joan</creatorcontrib><title>Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation</title><description>In recent years, domains such as natural language processing and image
recognition have popularized the paradigm of using large datasets to pretrain
representations that can be effectively transferred to downstream tasks. In
this work we evaluate how such a paradigm should be done in imitation learning,
where both pretraining and finetuning data are trajectories collected by
experts interacting with an unknown environment. Namely, we consider a setting
where the pretraining corpus consists of multitask demonstrations and the task
for each demonstration is set by an unobserved latent context variable. The
goal is to use the pretraining corpus to learn a low dimensional representation
of the high dimensional (e.g., visual) observation space which can be
transferred to a novel context for finetuning on a limited dataset of
demonstrations. Among a variety of possible pretraining objectives, we argue
that inverse dynamics modeling -- i.e., predicting an action given the
observations appearing before and after it in the demonstration -- is
well-suited to this setting. We provide empirical evidence of this claim
through evaluations on a variety of simulated visuomotor manipulation problems.
While previous work has attempted various theoretical explanations regarding
the benefit of inverse dynamics modeling, we find that these arguments are
insufficient to explain the empirical advantages often observed in our
settings, and so we derive a novel analysis using a simple but general
environment model.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj1FLwzAUhfPig0x_gE_mD7QmaW6zPsrUWehwyN7LNbkdYWs6kjjcv3duwoED54MDH2MPUpR6DiCeMP74Y6kqAaWsmzncsnUbjhQT8ZdTwNHbxNeRckQffNjyjjCGxJfT5PgnHSIlChmzn87jMEW--t5nnzHteDv6K7hjNwPuE93_94xt3l43i_ei-1i2i-euwNpA4cjUxhFKMiAVaaWrRlirQEltnKoBjDSicVo6K79kA8aCcQNpoc7Bppqxx-vtRak_RD9iPPV_av1FrfoFlfZJbw</recordid><startdate>20230526</startdate><enddate>20230526</enddate><creator>Brandfonbrener, David</creator><creator>Nachum, Ofir</creator><creator>Bruna, Joan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230526</creationdate><title>Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation</title><author>Brandfonbrener, David ; Nachum, Ofir ; Bruna, Joan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-de767dea1e7512e424390cc252147d265571709d41dc1b1957c57dfe402402a93</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Brandfonbrener, David</creatorcontrib><creatorcontrib>Nachum, Ofir</creatorcontrib><creatorcontrib>Bruna, Joan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Brandfonbrener, David</au><au>Nachum, Ofir</au><au>Bruna, Joan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation</atitle><date>2023-05-26</date><risdate>2023</risdate><abstract>In recent years, domains such as natural language processing and image
recognition have popularized the paradigm of using large datasets to pretrain
representations that can be effectively transferred to downstream tasks. In
this work we evaluate how such a paradigm should be done in imitation learning,
where both pretraining and finetuning data are trajectories collected by
experts interacting with an unknown environment. Namely, we consider a setting
where the pretraining corpus consists of multitask demonstrations and the task
for each demonstration is set by an unobserved latent context variable. The
goal is to use the pretraining corpus to learn a low dimensional representation
of the high dimensional (e.g., visual) observation space which can be
transferred to a novel context for finetuning on a limited dataset of
demonstrations. Among a variety of possible pretraining objectives, we argue
that inverse dynamics modeling -- i.e., predicting an action given the
observations appearing before and after it in the demonstration -- is
well-suited to this setting. We provide empirical evidence of this claim
through evaluations on a variety of simulated visuomotor manipulation problems.
While previous work has attempted various theoretical explanations regarding
the benefit of inverse dynamics modeling, we find that these arguments are
insufficient to explain the empirical advantages often observed in our
settings, and so we derive a novel analysis using a simple but general
environment model.</abstract><doi>10.48550/arxiv.2305.16985</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2305.16985 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2305_16985 |
source | arXiv.org |
subjects | Computer Science - Learning |
title | Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T10%3A15%3A00IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Inverse%20Dynamics%20Pretraining%20Learns%20Good%20Representations%20for%20Multitask%20Imitation&rft.au=Brandfonbrener,%20David&rft.date=2023-05-26&rft_id=info:doi/10.48550/arxiv.2305.16985&rft_dat=%3Carxiv_GOX%3E2305_16985%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |