Audio Driven
Audio-driven research focuses on understanding and generating audio signals, often in conjunction with other modalities like text and video. Current efforts concentrate on developing robust models for tasks such as audio-visual representation learning, talking head synthesis (using diffusion models and autoencoders), and audio-to-text/text-to-audio generation (leveraging large language models and neural codecs). These advancements have significant implications for various fields, including film-making, virtual reality, assistive technologies, and multimedia forensics, by enabling more realistic and interactive audio-visual experiences and improving analysis of audio-visual data.
Papers
September 6, 2022
August 24, 2022
August 23, 2022
August 4, 2022
June 29, 2022
June 28, 2022
June 26, 2022
May 27, 2022
April 27, 2022
April 15, 2022
March 21, 2022
March 18, 2022
February 25, 2022
February 21, 2022
February 15, 2022
January 11, 2022
December 16, 2021
November 25, 2021
November 8, 2021