MoCaNet: Motion Retargeting in-the-wild via Canonicalization Networks
We present a novel framework that brings the 3D motion retargeting task from controlled environments to in-the-wild scenarios. In particular, our method is capable of retargeting body motion from a character in a 2D monocular video to a 3D character without using any motion capture system or 3D reco...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We present a novel framework that brings the 3D motion retargeting task from
controlled environments to in-the-wild scenarios. In particular, our method is
capable of retargeting body motion from a character in a 2D monocular video to
a 3D character without using any motion capture system or 3D reconstruction
procedure. It is designed to leverage massive online videos for unsupervised
training, needless of 3D annotations or motion-body pairing information. The
proposed method is built upon two novel canonicalization operations, structure
canonicalization and view canonicalization. Trained with the canonicalization
operations and the derived regularizations, our method learns to factorize a
skeleton sequence into three independent semantic subspaces, i.e., motion,
structure, and view angle. The disentangled representation enables motion
retargeting from 2D to 3D with high precision. Our method achieves superior
performance on motion transfer benchmarks with large body variations and
challenging actions. Notably, the canonicalized skeleton sequence could serve
as a disentangled and interpretable representation of human motion that
benefits action analysis and motion retrieval. |
---|---|
DOI: | 10.48550/arxiv.2112.10082 |