Traditional RNNs
Traditional recurrent neural networks (RNNs) process sequential data by maintaining an internal state that is updated at each time step, aiming to capture temporal dependencies. Current research focuses on improving RNN learnability, particularly for long sequences, exploring architectures like LSTMs and GRUs, and investigating novel designs such as Mamba and RWKV that address limitations in computational efficiency and long-term memory. These efforts are driven by the need for more robust and efficient sequence models with improved interpretability, impacting diverse fields including time series forecasting, natural language processing, and image analysis.
Papers
September 26, 2023
September 18, 2023
August 28, 2023
August 22, 2023
June 24, 2023
June 1, 2023
May 22, 2023
May 18, 2023
April 7, 2023
March 12, 2023
February 21, 2023
February 6, 2023
January 26, 2023
January 11, 2023
December 10, 2022
December 9, 2022
November 29, 2022
November 6, 2022