Neural Transducer
Neural transducers are end-to-end sequence-to-sequence models primarily used for automatic speech recognition (ASR) and increasingly for tasks like speech translation and text-to-speech. Current research focuses on improving their efficiency and accuracy through techniques like self-supervised pre-training, factorized model architectures (e.g., incorporating separate language models), and novel training strategies such as sequence discriminative training and blank-symbol regularization. These advancements aim to create faster, more accurate, and resource-efficient models with applications in various fields, including improved voice assistants, real-time translation systems, and more robust speech processing in noisy environments.
Papers
September 30, 2024
August 6, 2024
June 6, 2024
April 2, 2024
January 3, 2024
November 6, 2023
October 13, 2023
October 11, 2023
September 25, 2023
September 18, 2023
September 15, 2023
September 14, 2023
May 26, 2023
May 19, 2023
May 9, 2023
March 30, 2023
March 20, 2023
February 27, 2023
December 7, 2022