Constrained Reinforcement Learning for Dexterous Manipulation
Existing learning approaches to dexterous manipulation use demonstrations or interactions with the environment to train black-box neural networks that provide little control over how the robot learns the skills or how it would perform post training. These approaches pose significant challenges when...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Jain, Abhineet Kolb, Jack Ravichandar, Harish |
description | Existing learning approaches to dexterous manipulation use demonstrations or
interactions with the environment to train black-box neural networks that
provide little control over how the robot learns the skills or how it would
perform post training. These approaches pose significant challenges when
implemented on physical platforms given that, during initial stages of
training, the robot's behavior could be erratic and potentially harmful to its
own hardware, the environment, or any humans in the vicinity. A potential way
to address these limitations is to add constraints during learning that
restrict and guide the robot's behavior during training as well as roll outs.
Inspired by the success of constrained approaches in other domains, we
investigate the effects of adding position-based constraints to a 24-DOF robot
hand learning to perform object relocation using Constrained Policy
Optimization. We find that a simple geometric constraint can ensure the robot
learns to move towards the object sooner than without constraints. Further,
training with this constraint requires a similar number of samples as its
unconstrained counterpart to master the skill. These findings shed light on how
simple constraints can help robots achieve sensible and safe behavior quickly
and ease concerns surrounding hardware deployment. We also investigate the
effects of the strictness of these constraints and report findings that provide
insights into how different degrees of strictness affect learning outcomes. Our
code is available at
https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation. |
doi_str_mv | 10.48550/arxiv.2301.09766 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2301_09766</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2301_09766</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-82d96904038a674dabb28bb19504ca587fbb6afa6591f851aa97f061019aa2ef3</originalsourceid><addsrcrecordid>eNotj7FuwjAURb0wVNAP6IR_IOlzEjv2wIBSKJWCkBB79EyekSVwkBMQ_XsodLq6dzi6h7EPAWmhpYRPjDd_TbMcRAqmVOqNzaou9ENEH6jlW_LBdXFPJwoDrwlj8OHAHxP_ottAsbv0fI3Bny9HHHwXJmzk8NjT-3-O2W652FWrpN58_1TzOkFVqkRnrVEGCsj1oxctWptpa4WRUOxR6tJZq9ChkkY4LQWiKR0oAcIgZuTyMZu-sM__zTn6E8bf5s-jeXrkd40yQ2A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Constrained Reinforcement Learning for Dexterous Manipulation</title><source>arXiv.org</source><creator>Jain, Abhineet ; Kolb, Jack ; Ravichandar, Harish</creator><creatorcontrib>Jain, Abhineet ; Kolb, Jack ; Ravichandar, Harish</creatorcontrib><description>Existing learning approaches to dexterous manipulation use demonstrations or
interactions with the environment to train black-box neural networks that
provide little control over how the robot learns the skills or how it would
perform post training. These approaches pose significant challenges when
implemented on physical platforms given that, during initial stages of
training, the robot's behavior could be erratic and potentially harmful to its
own hardware, the environment, or any humans in the vicinity. A potential way
to address these limitations is to add constraints during learning that
restrict and guide the robot's behavior during training as well as roll outs.
Inspired by the success of constrained approaches in other domains, we
investigate the effects of adding position-based constraints to a 24-DOF robot
hand learning to perform object relocation using Constrained Policy
Optimization. We find that a simple geometric constraint can ensure the robot
learns to move towards the object sooner than without constraints. Further,
training with this constraint requires a similar number of samples as its
unconstrained counterpart to master the skill. These findings shed light on how
simple constraints can help robots achieve sensible and safe behavior quickly
and ease concerns surrounding hardware deployment. We also investigate the
effects of the strictness of these constraints and report findings that provide
insights into how different degrees of strictness affect learning outcomes. Our
code is available at
https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.</description><identifier>DOI: 10.48550/arxiv.2301.09766</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Computer Science - Robotics</subject><creationdate>2023-01</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2301.09766$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2301.09766$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Jain, Abhineet</creatorcontrib><creatorcontrib>Kolb, Jack</creatorcontrib><creatorcontrib>Ravichandar, Harish</creatorcontrib><title>Constrained Reinforcement Learning for Dexterous Manipulation</title><description>Existing learning approaches to dexterous manipulation use demonstrations or
interactions with the environment to train black-box neural networks that
provide little control over how the robot learns the skills or how it would
perform post training. These approaches pose significant challenges when
implemented on physical platforms given that, during initial stages of
training, the robot's behavior could be erratic and potentially harmful to its
own hardware, the environment, or any humans in the vicinity. A potential way
to address these limitations is to add constraints during learning that
restrict and guide the robot's behavior during training as well as roll outs.
Inspired by the success of constrained approaches in other domains, we
investigate the effects of adding position-based constraints to a 24-DOF robot
hand learning to perform object relocation using Constrained Policy
Optimization. We find that a simple geometric constraint can ensure the robot
learns to move towards the object sooner than without constraints. Further,
training with this constraint requires a similar number of samples as its
unconstrained counterpart to master the skill. These findings shed light on how
simple constraints can help robots achieve sensible and safe behavior quickly
and ease concerns surrounding hardware deployment. We also investigate the
effects of the strictness of these constraints and report findings that provide
insights into how different degrees of strictness affect learning outcomes. Our
code is available at
https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Robotics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj7FuwjAURb0wVNAP6IR_IOlzEjv2wIBSKJWCkBB79EyekSVwkBMQ_XsodLq6dzi6h7EPAWmhpYRPjDd_TbMcRAqmVOqNzaou9ENEH6jlW_LBdXFPJwoDrwlj8OHAHxP_ottAsbv0fI3Bny9HHHwXJmzk8NjT-3-O2W652FWrpN58_1TzOkFVqkRnrVEGCsj1oxctWptpa4WRUOxR6tJZq9ChkkY4LQWiKR0oAcIgZuTyMZu-sM__zTn6E8bf5s-jeXrkd40yQ2A</recordid><startdate>20230123</startdate><enddate>20230123</enddate><creator>Jain, Abhineet</creator><creator>Kolb, Jack</creator><creator>Ravichandar, Harish</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230123</creationdate><title>Constrained Reinforcement Learning for Dexterous Manipulation</title><author>Jain, Abhineet ; Kolb, Jack ; Ravichandar, Harish</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-82d96904038a674dabb28bb19504ca587fbb6afa6591f851aa97f061019aa2ef3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Robotics</topic><toplevel>online_resources</toplevel><creatorcontrib>Jain, Abhineet</creatorcontrib><creatorcontrib>Kolb, Jack</creatorcontrib><creatorcontrib>Ravichandar, Harish</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Jain, Abhineet</au><au>Kolb, Jack</au><au>Ravichandar, Harish</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Constrained Reinforcement Learning for Dexterous Manipulation</atitle><date>2023-01-23</date><risdate>2023</risdate><abstract>Existing learning approaches to dexterous manipulation use demonstrations or
interactions with the environment to train black-box neural networks that
provide little control over how the robot learns the skills or how it would
perform post training. These approaches pose significant challenges when
implemented on physical platforms given that, during initial stages of
training, the robot's behavior could be erratic and potentially harmful to its
own hardware, the environment, or any humans in the vicinity. A potential way
to address these limitations is to add constraints during learning that
restrict and guide the robot's behavior during training as well as roll outs.
Inspired by the success of constrained approaches in other domains, we
investigate the effects of adding position-based constraints to a 24-DOF robot
hand learning to perform object relocation using Constrained Policy
Optimization. We find that a simple geometric constraint can ensure the robot
learns to move towards the object sooner than without constraints. Further,
training with this constraint requires a similar number of samples as its
unconstrained counterpart to master the skill. These findings shed light on how
simple constraints can help robots achieve sensible and safe behavior quickly
and ease concerns surrounding hardware deployment. We also investigate the
effects of the strictness of these constraints and report findings that provide
insights into how different degrees of strictness affect learning outcomes. Our
code is available at
https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.</abstract><doi>10.48550/arxiv.2301.09766</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2301.09766 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2301_09766 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning Computer Science - Robotics |
title | Constrained Reinforcement Learning for Dexterous Manipulation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T07%3A57%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Constrained%20Reinforcement%20Learning%20for%20Dexterous%20Manipulation&rft.au=Jain,%20Abhineet&rft.date=2023-01-23&rft_id=info:doi/10.48550/arxiv.2301.09766&rft_dat=%3Carxiv_GOX%3E2301_09766%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |