Face Video Generation
Face video generation aims to synthesize realistic and expressive talking head videos from audio input, often leveraging a reference image for identity preservation. Current research emphasizes improving lip synchronization, enhancing visual fidelity, and addressing challenges like motion jitters and expression control through techniques such as diffusion models, 3D Morphable Models (3DMMs), and StyleGANs, often incorporating intermediate representations like facial landmarks. These advancements have implications for various applications, including virtual assistants, video conferencing, and film production, by enabling more natural and engaging human-computer interaction and creative content generation.
Papers
October 14, 2024
September 23, 2024
August 10, 2024
July 12, 2024
February 28, 2024
September 17, 2022
August 29, 2022
August 16, 2022
January 16, 2022