Recurrent Network
Recurrent networks are neural network architectures designed to process sequential data by maintaining an internal state that evolves over time, enabling them to capture temporal dependencies. Current research focuses on improving their efficiency and scalability, particularly through novel architectures like state-space models and modifications to classic RNNs (LSTMs, GRUs) that enable parallel training. This renewed interest stems from limitations in transformer models for long sequences and a desire for more biologically plausible learning algorithms, leading to advancements in areas like online learning and applications in diverse fields such as recommender systems, medical image registration, and robotics.
Papers
Beyond Geometry: Comparing the Temporal Structure of Computation in Neural Circuits with Dynamical Similarity Analysis
Mitchell Ostrow, Adam Eisen, Leo Kozachkov, Ila Fiete
Reducing Computational Costs in Sentiment Analysis: Tensorized Recurrent Networks vs. Recurrent Networks
Gabriel Lopez, Anna Nguyen, Joe Kaul