Constrained Reinforcement Learning for Dexterous Manipulation

Existing learning approaches to dexterous manipulation use demonstrations or interactions with the environment to train black-box neural networks that provide little control over how the robot learns the skills or how it would perform post training. These approaches pose significant challenges when...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Jain, Abhineet, Kolb, Jack, Ravichandar, Harish
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Jain, Abhineet
Kolb, Jack
Ravichandar, Harish
description Existing learning approaches to dexterous manipulation use demonstrations or interactions with the environment to train black-box neural networks that provide little control over how the robot learns the skills or how it would perform post training. These approaches pose significant challenges when implemented on physical platforms given that, during initial stages of training, the robot's behavior could be erratic and potentially harmful to its own hardware, the environment, or any humans in the vicinity. A potential way to address these limitations is to add constraints during learning that restrict and guide the robot's behavior during training as well as roll outs. Inspired by the success of constrained approaches in other domains, we investigate the effects of adding position-based constraints to a 24-DOF robot hand learning to perform object relocation using Constrained Policy Optimization. We find that a simple geometric constraint can ensure the robot learns to move towards the object sooner than without constraints. Further, training with this constraint requires a similar number of samples as its unconstrained counterpart to master the skill. These findings shed light on how simple constraints can help robots achieve sensible and safe behavior quickly and ease concerns surrounding hardware deployment. We also investigate the effects of the strictness of these constraints and report findings that provide insights into how different degrees of strictness affect learning outcomes. Our code is available at https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.
doi_str_mv 10.48550/arxiv.2301.09766
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2301_09766</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2301_09766</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-82d96904038a674dabb28bb19504ca587fbb6afa6591f851aa97f061019aa2ef3</originalsourceid><addsrcrecordid>eNotj7FuwjAURb0wVNAP6IR_IOlzEjv2wIBSKJWCkBB79EyekSVwkBMQ_XsodLq6dzi6h7EPAWmhpYRPjDd_TbMcRAqmVOqNzaou9ENEH6jlW_LBdXFPJwoDrwlj8OHAHxP_ottAsbv0fI3Bny9HHHwXJmzk8NjT-3-O2W652FWrpN58_1TzOkFVqkRnrVEGCsj1oxctWptpa4WRUOxR6tJZq9ChkkY4LQWiKR0oAcIgZuTyMZu-sM__zTn6E8bf5s-jeXrkd40yQ2A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Constrained Reinforcement Learning for Dexterous Manipulation</title><source>arXiv.org</source><creator>Jain, Abhineet ; Kolb, Jack ; Ravichandar, Harish</creator><creatorcontrib>Jain, Abhineet ; Kolb, Jack ; Ravichandar, Harish</creatorcontrib><description>Existing learning approaches to dexterous manipulation use demonstrations or interactions with the environment to train black-box neural networks that provide little control over how the robot learns the skills or how it would perform post training. These approaches pose significant challenges when implemented on physical platforms given that, during initial stages of training, the robot's behavior could be erratic and potentially harmful to its own hardware, the environment, or any humans in the vicinity. A potential way to address these limitations is to add constraints during learning that restrict and guide the robot's behavior during training as well as roll outs. Inspired by the success of constrained approaches in other domains, we investigate the effects of adding position-based constraints to a 24-DOF robot hand learning to perform object relocation using Constrained Policy Optimization. We find that a simple geometric constraint can ensure the robot learns to move towards the object sooner than without constraints. Further, training with this constraint requires a similar number of samples as its unconstrained counterpart to master the skill. These findings shed light on how simple constraints can help robots achieve sensible and safe behavior quickly and ease concerns surrounding hardware deployment. We also investigate the effects of the strictness of these constraints and report findings that provide insights into how different degrees of strictness affect learning outcomes. Our code is available at https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.</description><identifier>DOI: 10.48550/arxiv.2301.09766</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Computer Science - Robotics</subject><creationdate>2023-01</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2301.09766$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2301.09766$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Jain, Abhineet</creatorcontrib><creatorcontrib>Kolb, Jack</creatorcontrib><creatorcontrib>Ravichandar, Harish</creatorcontrib><title>Constrained Reinforcement Learning for Dexterous Manipulation</title><description>Existing learning approaches to dexterous manipulation use demonstrations or interactions with the environment to train black-box neural networks that provide little control over how the robot learns the skills or how it would perform post training. These approaches pose significant challenges when implemented on physical platforms given that, during initial stages of training, the robot's behavior could be erratic and potentially harmful to its own hardware, the environment, or any humans in the vicinity. A potential way to address these limitations is to add constraints during learning that restrict and guide the robot's behavior during training as well as roll outs. Inspired by the success of constrained approaches in other domains, we investigate the effects of adding position-based constraints to a 24-DOF robot hand learning to perform object relocation using Constrained Policy Optimization. We find that a simple geometric constraint can ensure the robot learns to move towards the object sooner than without constraints. Further, training with this constraint requires a similar number of samples as its unconstrained counterpart to master the skill. These findings shed light on how simple constraints can help robots achieve sensible and safe behavior quickly and ease concerns surrounding hardware deployment. We also investigate the effects of the strictness of these constraints and report findings that provide insights into how different degrees of strictness affect learning outcomes. Our code is available at https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Robotics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj7FuwjAURb0wVNAP6IR_IOlzEjv2wIBSKJWCkBB79EyekSVwkBMQ_XsodLq6dzi6h7EPAWmhpYRPjDd_TbMcRAqmVOqNzaou9ENEH6jlW_LBdXFPJwoDrwlj8OHAHxP_ottAsbv0fI3Bny9HHHwXJmzk8NjT-3-O2W652FWrpN58_1TzOkFVqkRnrVEGCsj1oxctWptpa4WRUOxR6tJZq9ChkkY4LQWiKR0oAcIgZuTyMZu-sM__zTn6E8bf5s-jeXrkd40yQ2A</recordid><startdate>20230123</startdate><enddate>20230123</enddate><creator>Jain, Abhineet</creator><creator>Kolb, Jack</creator><creator>Ravichandar, Harish</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230123</creationdate><title>Constrained Reinforcement Learning for Dexterous Manipulation</title><author>Jain, Abhineet ; Kolb, Jack ; Ravichandar, Harish</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-82d96904038a674dabb28bb19504ca587fbb6afa6591f851aa97f061019aa2ef3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Robotics</topic><toplevel>online_resources</toplevel><creatorcontrib>Jain, Abhineet</creatorcontrib><creatorcontrib>Kolb, Jack</creatorcontrib><creatorcontrib>Ravichandar, Harish</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Jain, Abhineet</au><au>Kolb, Jack</au><au>Ravichandar, Harish</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Constrained Reinforcement Learning for Dexterous Manipulation</atitle><date>2023-01-23</date><risdate>2023</risdate><abstract>Existing learning approaches to dexterous manipulation use demonstrations or interactions with the environment to train black-box neural networks that provide little control over how the robot learns the skills or how it would perform post training. These approaches pose significant challenges when implemented on physical platforms given that, during initial stages of training, the robot's behavior could be erratic and potentially harmful to its own hardware, the environment, or any humans in the vicinity. A potential way to address these limitations is to add constraints during learning that restrict and guide the robot's behavior during training as well as roll outs. Inspired by the success of constrained approaches in other domains, we investigate the effects of adding position-based constraints to a 24-DOF robot hand learning to perform object relocation using Constrained Policy Optimization. We find that a simple geometric constraint can ensure the robot learns to move towards the object sooner than without constraints. Further, training with this constraint requires a similar number of samples as its unconstrained counterpart to master the skill. These findings shed light on how simple constraints can help robots achieve sensible and safe behavior quickly and ease concerns surrounding hardware deployment. We also investigate the effects of the strictness of these constraints and report findings that provide insights into how different degrees of strictness affect learning outcomes. Our code is available at https://github.com/GT-STAR-Lab/constrained-rl-dexterous-manipulation.</abstract><doi>10.48550/arxiv.2301.09766</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2301.09766
ispartof
issn
language eng
recordid cdi_arxiv_primary_2301_09766
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
Computer Science - Robotics
title Constrained Reinforcement Learning for Dexterous Manipulation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T07%3A57%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Constrained%20Reinforcement%20Learning%20for%20Dexterous%20Manipulation&rft.au=Jain,%20Abhineet&rft.date=2023-01-23&rft_id=info:doi/10.48550/arxiv.2301.09766&rft_dat=%3Carxiv_GOX%3E2301_09766%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true