UniFaceGAN: A Unified Framework for Temporally Consistent Facial Video Editing

Meng Cao, Haozhi Huang, Hao Wang, Xuan Wang, Li Shen, Sheng Wang, Linchao Bao*, Zhifeng Li, Jiebo Luo

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Recent research has witnessed advances in facial image editing tasks including face swapping and face reenactment. However, these methods are confined to dealing with one specific task at a time. In addition, for video facial editing, previous methods either simply apply transformations frame by frame or utilize multiple frames in a concatenated or iterative fashion, which leads to noticeable visual flickers. In this paper, we propose a unified temporally consistent facial video editing framework termed UniFaceGAN. Based on a 3D reconstruction model and a simple yet efficient dynamic training sample selection mechanism, our framework is designed to handle face swapping and face reenactment simultaneously. To enforce the temporal consistency, a novel 3D temporal loss constraint is introduced based on the barycentric coordinate interpolation. Besides, we propose a region-aware conditional normalization layer to replace the traditional AdaIN or SPADE to synthesize more context-harmonious results. Compared with the state-of-the-art facial image editing methods, our framework generates video portraits that are more photo-realistic and temporally smooth.
Original languageEnglish
Article number9464699
Pages (from-to)6107-6116
Number of pages10
JournalIEEE Transactions on Image Processing
Volume30
Early online date24 Jun 2021
DOIs
Publication statusPublished - 7 Jul 2021

Keywords

  • Faces
  • Three-dimensional displays
  • Training
  • Task analysis
  • Image reconstruction
  • Optical losses
  • Solid modeling

Fingerprint

Dive into the research topics of 'UniFaceGAN: A Unified Framework for Temporally Consistent Facial Video Editing'. Together they form a unique fingerprint.

Cite this