PKU-DyMVHumans: A Multi-View Video Benchmark for High-Fidelity Dynamic Human Modeling

Xiaoyun Zheng, Liwei Liao, Xufeng Li, Jianbo Jiao, Rongjie Wang, Feng Gao, Shiqi Wang, Ronggang Wang

Research output: Working paper/PreprintPreprint

35 Downloads (Pure)

Abstract

High-quality human reconstruction and photo-realistic rendering of a dynamic scene is a long-standing problem in computer vision and graphics. Despite considerable efforts invested in developing various capture systems and reconstruction algorithms, recent advancements still struggle with loose or oversized clothing and overly complex poses. In part, this is due to the challenges of acquiring high-quality human datasets. To facilitate the development of these fields, in this paper, we present PKU-DyMVHumans, a versatile human-centric dataset for high-fidelity reconstruction and rendering of dynamic human scenarios from dense multi-view videos. It comprises 8.2 million frames captured by more than 56 synchronized cameras across diverse scenarios. These sequences comprise 32 human subjects across 45 different scenarios, each with a high-detailed appearance and realistic human motion. Inspired by recent advancements in neural radiance field (NeRF)-based scene representations, we carefully set up an off-the-shelf framework that is easy to provide those state-of-the-art NeRF-based implementations and benchmark on PKU-DyMVHumans dataset. It is paving the way for various applications like fine-grained foreground/background decomposition, high-quality human reconstruction and photo-realistic novel view synthesis of a dynamic scene. Extensive studies are performed on the benchmark, demonstrating new observations and challenges that emerge from using such high-fidelity dynamic data. The dataset is available at: https://pku-dymvhumans.github.io.
Original languageEnglish
PublisherarXiv
Number of pages24
DOIs
Publication statusPublished - 24 Mar 2024

Bibliographical note

Acknowledgments: This work is supported by Outstanding Talents Training Fund in Shenzhen, Shenzhen Science and Technology Program (RCJC20200714114435057, SGDX20211123144400001), National Natural Science Foundation of China (U21B2012), and Migu-PKU Meta Vision Technology Innovation Laboratory (R24115SG). Jianbo Jiao is supported by the Royal Society grants IES\R3\223050 and SIF\R1\231009.

Keywords

  • cs.CV

Fingerprint

Dive into the research topics of 'PKU-DyMVHumans: A Multi-View Video Benchmark for High-Fidelity Dynamic Human Modeling'. Together they form a unique fingerprint.

Cite this