Reducing Sim-to-Real Domain Gaps of Visual Sensors for Learning Environment-constrained Visuomotor Policy

Simulation engines enable safe training of robotic skills, but domain gaps between simulated and real sensors hinder deployment. However, existing pixel-level adaptation methods focus on visual realism of generating images over task-specific learning, causing texture leakage and elimination. In this...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE sensors journal 2024-12, p.1-1
Hauptverfasser: Jing, Xingshuo, Qian, Kun, Duan, Boyi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Simulation engines enable safe training of robotic skills, but domain gaps between simulated and real sensors hinder deployment. However, existing pixel-level adaptation methods focus on visual realism of generating images over task-specific learning, causing texture leakage and elimination. In this paper, we introduce LCTGAN, an novel unsupervised domain transfer network, with a correlative attention mechanism and a mask-level Q-value mapping consistency to enhance task-awareness and bridge the domain gap of visual sensors in pixel-level perceptual manipulations. We also propose a Q-learning-based visuomotor policy to handle cluttered scenarios where objects may lack directly graspable configurations, which learns the synergies of three actions while considering environmental constraints. We further integrate LCTGAN into the learned policy to facilitate zero-shot sim-to-real policy transfer. Extensive experimental results validate the zero-shot sim-to-real generalization of our proposed visuomotor policy when deployed on a real robot. The supplementary video is available at: https://youtu.be/B6nODKkhzSw.
ISSN:1530-437X
1558-1748
DOI:10.1109/JSEN.2024.3522245