Audio Driven
Audio-driven research focuses on understanding and generating audio signals, often in conjunction with other modalities like text and video. Current efforts concentrate on developing robust models for tasks such as audio-visual representation learning, talking head synthesis (using diffusion models and autoencoders), and audio-to-text/text-to-audio generation (leveraging large language models and neural codecs). These advancements have significant implications for various fields, including film-making, virtual reality, assistive technologies, and multimedia forensics, by enabling more realistic and interactive audio-visual experiences and improving analysis of audio-visual data.
Papers
March 1, 2023
February 26, 2023
February 24, 2023
February 14, 2023
February 8, 2023
January 21, 2023
January 17, 2023
January 15, 2023
December 9, 2022
December 2, 2022
November 21, 2022
November 20, 2022
November 17, 2022
October 27, 2022
October 5, 2022
September 30, 2022
September 26, 2022
September 9, 2022