Talking Face Generation
Talking face generation aims to synthesize realistic and synchronized videos of a person speaking, given only an audio recording and potentially a single image of their face. Current research focuses on improving lip synchronization accuracy, generating natural head movements and facial expressions (including emotions), and enhancing the overall realism and visual quality of the generated videos, often employing diffusion models, GANs, and transformer-based architectures. These advancements have implications for various fields, including virtual assistants, video conferencing, and healthcare applications like creating personalized avatars for Alzheimer's patients.
Papers
October 9, 2024
May 23, 2024
May 12, 2024
May 7, 2024
March 4, 2024
February 25, 2024
January 16, 2024
December 12, 2023
November 20, 2023
April 2, 2023
March 30, 2023
January 16, 2023
January 6, 2023
August 29, 2022
June 29, 2022