Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled Representation
The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 6040-6049 We present a lightweight solution to recover 3D pose from multi-view images captured with spatially calibrated cameras. Building upon recent advances in interpretable representation learning, we exploit 3D...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The IEEE/CVF Conference on Computer Vision and Pattern Recognition
(CVPR), 2020, pp. 6040-6049 We present a lightweight solution to recover 3D pose from multi-view images
captured with spatially calibrated cameras. Building upon recent advances in
interpretable representation learning, we exploit 3D geometry to fuse input
images into a unified latent representation of pose, which is disentangled from
camera view-points. This allows us to reason effectively about 3D pose across
different views without using compute-intensive volumetric grids. Our
architecture then conditions the learned representation on camera projection
operators to produce accurate per-view 2d detections, that can be simply lifted
to 3D via a differentiable Direct Linear Transform (DLT) layer. In order to do
it efficiently, we propose a novel implementation of DLT that is orders of
magnitude faster on GPU architectures than standard SVD-based triangulation
methods. We evaluate our approach on two large-scale human pose datasets (H36M
and Total Capture): our method outperforms or performs comparably to the
state-of-the-art volumetric methods, while, unlike them, yielding real-time
performance. |
---|---|
DOI: | 10.48550/arxiv.2004.02186 |