RNN T Model
Recurrent Neural Network Transducers (RNN-Ts) are a prominent architecture for automatic speech recognition (ASR), aiming to improve accuracy and efficiency compared to traditional methods. Current research focuses on optimizing RNN-T inference speed through techniques like GPU-accelerated decoding and novel architectures such as CIF-T, which aim to reduce computational redundancy. These advancements, along with explorations of knowledge distillation and efficient quantization methods, are driving improvements in both the accuracy and real-time performance of RNN-T models for various ASR applications.
Papers
June 6, 2024
March 13, 2024
July 26, 2023
March 10, 2023
June 29, 2022
June 16, 2022
February 26, 2022
November 19, 2021