Mt RNN
Recurrent Neural Network Transducers (RNN-Ts) are a prominent architecture for end-to-end speech recognition, aiming to improve accuracy and efficiency in streaming applications. Current research focuses on optimizing RNN-T models through techniques like efficient decoding algorithms (e.g., greedy decoding, beam search), model compression (e.g., weight binarization), and addressing challenges such as training data imperfections and robustness to adversarial attacks. These advancements are significant for improving the speed, memory efficiency, and accuracy of speech recognition systems, impacting both research in automatic speech recognition and the development of real-world applications like voice assistants and voice search.
Papers
September 16, 2024
August 22, 2024
August 19, 2024
June 26, 2024
June 6, 2024
June 5, 2024
September 22, 2023
September 5, 2023
July 17, 2023
July 11, 2023
June 24, 2023
May 31, 2023
May 12, 2023
April 17, 2023
March 18, 2023
February 28, 2023
December 21, 2022
November 4, 2022
November 2, 2022