Talking Face
Talking face generation aims to synthesize realistic and expressive video of a person speaking, driven by audio input or text. Current research heavily focuses on improving the realism and emotional expressiveness of these synthetic videos, employing techniques like diffusion models, generative adversarial networks (GANs), and variational autoencoders (VAEs) to achieve accurate lip-sync and nuanced facial expressions, often incorporating disentangled representations of identity, emotion, and motion. This field is significant for its applications in animation, virtual reality, and communication technologies, as well as its potential to advance research in computer vision, machine learning, and human-computer interaction.
Papers
August 12, 2024
June 26, 2024
May 16, 2024
May 6, 2024
March 11, 2024
March 4, 2024
February 25, 2024
December 21, 2023
December 15, 2023
December 5, 2023
November 15, 2023
September 14, 2023
April 18, 2023
April 6, 2023
March 24, 2023
March 9, 2023
May 30, 2022