Recurrent Neural Network
Recurrent Neural Networks (RNNs) are a class of neural networks designed to process sequential data by maintaining an internal state that is updated over time. Current research focuses on improving RNN efficiency and stability, exploring variations like LSTMs and GRUs, and investigating their application in diverse fields such as time series forecasting, natural language processing, and dynamical systems modeling. This includes developing novel architectures like selective state space models for improved memory efficiency and exploring the use of RNNs in conjunction with other architectures, such as transformers and convolutional neural networks. The resulting advancements have significant implications for various applications requiring sequential data processing, offering improved accuracy, efficiency, and interpretability.
Papers - Page 2
Discovering Chunks in Neural Embeddings for Interpretability
Shuchen Wu, Stephan Alaniz, Eric Schulz, Zeynep AkataAccelerating Linear Recurrent Neural Networks for the Edge with Unstructured Sparsity
Alessandro Pierro, Steven Abreu, Jonathan Timcheck, Philipp Stratmann, Andreas Wild, Sumit Bam Shrestha
Predicting concentration levels of air pollutants by transfer learning and recurrent neural network
Iat Hang Fong, Tengyue Li, Simon Fong, Raymond K. Wong, Antonio J. Tallón-BallesterosBARNN: A Bayesian Autoregressive and Recurrent Neural Network
Dario Coscia, Max Welling, Nicola Demo, Gianluigi Rozza