Non Speech Audio
Non-speech audio research focuses on extracting meaningful information and patterns from audio signals excluding human speech, aiming to understand and utilize diverse acoustic phenomena. Current research emphasizes developing robust models, often based on transformer architectures and diffusion models, for tasks like audio editing, generation (e.g., from video), and classification across various domains (music, environmental sounds). This field is significant for advancing audio representation learning, enabling applications such as privacy-preserving crowd analysis, improved audio-visual systems, and more sophisticated audio-based content analysis for tasks like hate speech detection.
Papers
October 31, 2024
July 10, 2024
May 11, 2024
May 2, 2024
March 2, 2024
February 12, 2024
January 19, 2024
November 12, 2023
November 2, 2023
September 19, 2023
August 20, 2023
May 9, 2023
April 19, 2023
March 14, 2023
January 31, 2023
November 1, 2022
October 13, 2022
February 17, 2022