CORN: Contact-based Object Representation for Nonprehensile Manipulation of General Unseen Objects
Nonprehensile manipulation is essential for manipulating objects that are too thin, large, or otherwise ungraspable in the wild. To sidestep the difficulty of contact modeling in conventional modeling-based approaches, reinforcement learning (RL) has recently emerged as a promising alternative. Howe...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Nonprehensile manipulation is essential for manipulating objects that are too
thin, large, or otherwise ungraspable in the wild. To sidestep the difficulty
of contact modeling in conventional modeling-based approaches, reinforcement
learning (RL) has recently emerged as a promising alternative. However,
previous RL approaches either lack the ability to generalize over diverse
object shapes, or use simple action primitives that limit the diversity of
robot motions. Furthermore, using RL over diverse object geometry is
challenging due to the high cost of training a policy that takes in
high-dimensional sensory inputs. We propose a novel contact-based object
representation and pretraining pipeline to tackle this. To enable massively
parallel training, we leverage a lightweight patch-based transformer
architecture for our encoder that processes point clouds, thus scaling our
training across thousands of environments. Compared to learning from scratch,
or other shape representation baselines, our representation facilitates both
time- and data-efficient learning. We validate the efficacy of our overall
system by zero-shot transferring the trained policy to novel real-world
objects. Code and videos are available at
https://sites.google.com/view/contact-non-prehensile. |
---|---|
DOI: | 10.48550/arxiv.2403.10760 |