Neural Sequence Model
Neural sequence models aim to process and generate sequential data, such as text or music, by learning patterns and relationships within the sequences. Current research focuses on understanding how these models, particularly transformers and recurrent neural networks, achieve in-context learning and compositional generalization, investigating their internal mechanisms and exploring algorithmic interpretations of their behavior. This research is significant for advancing our understanding of artificial intelligence and has implications for various applications, including natural language processing, machine translation, and symbolic music understanding, by improving model performance and robustness.
Papers
June 17, 2024
January 23, 2024
December 13, 2023
November 14, 2023
August 24, 2023
June 7, 2023
June 5, 2023
May 30, 2023
January 18, 2023
November 28, 2022
October 18, 2022
June 5, 2022
April 1, 2022
March 14, 2022
January 30, 2022
December 16, 2021
November 28, 2021