Traditional RNNs
Traditional recurrent neural networks (RNNs) process sequential data by maintaining an internal state that is updated at each time step, aiming to capture temporal dependencies. Current research focuses on improving RNN learnability, particularly for long sequences, exploring architectures like LSTMs and GRUs, and investigating novel designs such as Mamba and RWKV that address limitations in computational efficiency and long-term memory. These efforts are driven by the need for more robust and efficient sequence models with improved interpretability, impacting diverse fields including time series forecasting, natural language processing, and image analysis.
Papers
Recurrent Neural Networks for Dynamical Systems: Applications to Ordinary Differential Equations, Collective Motion, and Hydrological Modeling
Yonggi Park, Kelum Gajamannage, Dilhani I. Jayathilake, Erik M. Bollt
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
Gianna Paulin, Francesco Conti, Lukas Cavigelli, Luca Benini