Contextual Reinforcement Learning of Visuo-tactile Multi-fingered Grasping Policies
Using simulation to train robot manipulation policies holds the promise of an almost unlimited amount of training data, generated safely out of harm's way. One of the key challenges of using simulation, to date, has been to bridge the reality gap, so that policies trained in simulation can be d...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Using simulation to train robot manipulation policies holds the promise of an
almost unlimited amount of training data, generated safely out of harm's way.
One of the key challenges of using simulation, to date, has been to bridge the
reality gap, so that policies trained in simulation can be deployed in the real
world. We explore the reality gap in the context of learning a contextual
policy for multi-fingered robotic grasping. We propose a Grasping Objects
Approach for Tactile (GOAT) robotic hands, learning to overcome the reality gap
problem. In our approach we use human hand motion demonstration to initialize
and reduce the search space for learning. We contextualize our policy with the
bounding cuboid dimensions of the object of interest, which allows the policy
to work on a more flexible representation than directly using an image or point
cloud. Leveraging fingertip touch sensors in the hand allows the policy to
overcome the reduction in geometric information introduced by the coarse
bounding box, as well as pose estimation uncertainty. We show our learned
policy successfully runs on a real robot without any fine tuning, thus bridging
the reality gap. |
---|---|
DOI: | 10.48550/arxiv.1911.09233 |