Self Supervised Speech Representation Model
Self-supervised speech representation models learn powerful representations of speech from vast amounts of unlabeled audio data, aiming to improve various downstream tasks like speech recognition and synthesis. Current research focuses on adapting these models for low-resource languages, enhancing noise robustness, and disentangling speaker characteristics from content for improved speech coding and paralinguistic analysis. These advancements are significantly impacting fields like hearing aid technology, cross-lingual speech processing, and animal vocalization analysis, demonstrating the broad applicability of these models.
Papers
September 30, 2024
July 18, 2024
April 29, 2024
January 10, 2024
October 17, 2023
May 18, 2023
May 9, 2023
April 24, 2023
November 22, 2022
October 21, 2022
December 27, 2021