Reducing Sim-to-Real Domain Gaps of Visual Sensors for Learning Environment-constrained Visuomotor Policy
Simulation engines enable safe training of robotic skills, but domain gaps between simulated and real sensors hinder deployment. However, existing pixel-level adaptation methods focus on visual realism of generating images over task-specific learning, causing texture leakage and elimination. In this...
Gespeichert in:
Veröffentlicht in: | IEEE sensors journal 2024-12, p.1-1 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Simulation engines enable safe training of robotic skills, but domain gaps between simulated and real sensors hinder deployment. However, existing pixel-level adaptation methods focus on visual realism of generating images over task-specific learning, causing texture leakage and elimination. In this paper, we introduce LCTGAN, an novel unsupervised domain transfer network, with a correlative attention mechanism and a mask-level Q-value mapping consistency to enhance task-awareness and bridge the domain gap of visual sensors in pixel-level perceptual manipulations. We also propose a Q-learning-based visuomotor policy to handle cluttered scenarios where objects may lack directly graspable configurations, which learns the synergies of three actions while considering environmental constraints. We further integrate LCTGAN into the learned policy to facilitate zero-shot sim-to-real policy transfer. Extensive experimental results validate the zero-shot sim-to-real generalization of our proposed visuomotor policy when deployed on a real robot. The supplementary video is available at: https://youtu.be/B6nODKkhzSw. |
---|---|
ISSN: | 1530-437X 1558-1748 |
DOI: | 10.1109/JSEN.2024.3522245 |