Language Modelling
Language modeling aims to build computational models that can understand and generate human language, primarily by predicting the probability of sequences of words. Current research focuses on improving model efficiency and generalization, exploring architectures like Transformers and LSTMs, and investigating techniques such as masked and causal language modeling, data augmentation, and contrastive learning to enhance performance across diverse tasks and languages. These advancements have significant implications for various applications, including machine translation, speech recognition, and question answering, as well as for fundamental research into language understanding and representation.
Papers
January 10, 2025
December 21, 2024
December 16, 2024
October 29, 2024
September 9, 2024
July 18, 2024
June 17, 2024
June 10, 2024
June 4, 2024
May 21, 2024
May 15, 2024
April 29, 2024
April 9, 2024
February 21, 2024
February 4, 2024
January 6, 2024
December 6, 2023
December 2, 2023