Robust and Accurate Monocular Pose Tracking for Large Pose Shift
Tracking the pose of a specific rigid object from monocular sequences is a basic problem in computer vision. State-of-the-art methods assume motion continuity between two consecutive frames. However, drastic relative motion causes large inter-frame pose shifts, especially in applications such as rob...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on industrial electronics (1982) 2023-08, Vol.70 (8), p.1-10 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Tracking the pose of a specific rigid object from monocular sequences is a basic problem in computer vision. State-of-the-art methods assume motion continuity between two consecutive frames. However, drastic relative motion causes large inter-frame pose shifts, especially in applications such as robotic grasping, failed satellite maintenance and space debris removal. Large pose shifts interrupt the inter-frame motion continuity leading to tracking failure. In this paper, we propose a robust and accurate monocular pose tracking method for tracking objects with large pose shifts. Using an indexable sparse viewpoint model to represent the object 3D geometry, we propose establishing a transitional view, which is searched for in an efficient variable-step way, to recover motion continuity. Then, a region-based optimization algorithm is adopted to optimize the pose based on the transitional view. Finally, we use a single-rendering-based pose refinement process to achieve highly accurate pose results. The experiments on the region-based object tracking (RBOT) dataset, the modified RBOT dataset, the synthetic large pose shift sequences and real sequences demonstrated that the proposed method achieved superior performance to the state-of-the-art methods in tracking objects with large pose shifts. |
---|---|
ISSN: | 0278-0046 1557-9948 |
DOI: | 10.1109/TIE.2022.3217598 |