Face Generation
Face generation research aims to create realistic and controllable facial images and videos, often driven by audio or text input. Current efforts focus on improving the realism and expressiveness of generated faces, employing diverse model architectures such as diffusion models, neural radiance fields (NeRFs), and transformers, often incorporating techniques like disentangled representations and fine-grained control mechanisms (e.g., action units). This field is significant for its applications in entertainment, virtual reality, and communication technologies, while also presenting challenges and opportunities in areas like deepfake detection and ethical considerations surrounding synthetic media.
Papers
Gaussian Harmony: Attaining Fairness in Diffusion-based Face Generation Models
Basudha Pal, Arunkumar Kannan, Ram Prabhakar Kathirvel, Alice J. O'Toole, Rama Chellappa
Controllable 3D Face Generation with Conditional Style Code Diffusion
Xiaolong Shen, Jianxin Ma, Chang Zhou, Zongxin Yang
DREAM-Talk: Diffusion-based Realistic Emotional Audio-driven Method for Single Image Talking Face Generation
Chenxu Zhang, Chao Wang, Jianfeng Zhang, Hongyi Xu, Guoxian Song, You Xie, Linjie Luo, Yapeng Tian, Xiaohu Guo, Jiashi Feng