PMnet: learning of disentangled pose and movement for unsupervised motion retargeting
Research output: Chapter in Book/Report/Conference proceeding › Conference contribution
Colleges, School and Institutes
- Seoul National University, Seoul, Korea.
In this paper, we propose a deep learning framework for unsupervised motion retargeting. In contrast to the existing method, we decouple the motion retargeting process into two parts that explicitly learn poses and movements of a character. Here, the first part retargets the pose of the character at each frame, while the second part retargets the character’s overall movement. To realize these two processes, we develop a novel architecture referred to as the pose-movement network (PMnet), which separately learns frame-by-frame poses and overall movement. At each frame, to follow the pose of the input character, PMnet learns how to make the input pose first and then adjusts it to fit the target character’s kinematic configuration. To handle the overall movement, a normalizing process is introduced to make the overall movement invariant to the size of the character. Along with the normalizing process, PMnet regresses the overall movement to fit the target character. We then introduce a novel loss function that allows PMnet to properly retarget the poses and overall movement. The proposed method is verified via several self-comparisons and outperforms the state-of-the-art (sota) method by reducing the motion retargeting error (average joint position error) from 7.68 (sota) to 1.95 (ours).
|Title of host publication||Proceedings of the 30th British Machine Vision Conference (BMVC 2019)|
|Publication status||Published - 12 Sep 2019|
|Event||30th British Machine Vision Conference (BMVC 2019) - Cardiff, United Kingdom|
Duration: 9 Sep 2019 → 12 Sep 2019
|Conference||30th British Machine Vision Conference (BMVC 2019)|
|Period||9/09/19 → 12/09/19|