Face Generation
Face generation research aims to create realistic and controllable facial images and videos, often driven by audio or text input. Current efforts focus on improving the realism and expressiveness of generated faces, employing diverse model architectures such as diffusion models, neural radiance fields (NeRFs), and transformers, often incorporating techniques like disentangled representations and fine-grained control mechanisms (e.g., action units). This field is significant for its applications in entertainment, virtual reality, and communication technologies, while also presenting challenges and opportunities in areas like deepfake detection and ethical considerations surrounding synthetic media.
Papers
StyleHEAT: One-Shot High-Resolution Editable Talking Face Generation via Pre-trained StyleGAN
Fei Yin, Yong Zhang, Xiaodong Cun, Mingdeng Cao, Yanbo Fan, Xuan Wang, Qingyan Bai, Baoyuan Wu, Jue Wang, Yujiu Yang
Attention-Based Lip Audio-Visual Synthesis for Talking Face Generation in the Wild
Ganglai Wang, Peng Zhang, Lei Xie, Wei Huang, Yufei Zha