Language Modelling
Language modeling aims to build computational models that can understand and generate human language, primarily by predicting the probability of sequences of words. Current research focuses on improving model efficiency and generalization, exploring architectures like Transformers and LSTMs, and investigating techniques such as masked and causal language modeling, data augmentation, and contrastive learning to enhance performance across diverse tasks and languages. These advancements have significant implications for various applications, including machine translation, speech recognition, and question answering, as well as for fundamental research into language understanding and representation.
Papers
October 23, 2023
September 27, 2023
September 20, 2023
August 15, 2023
August 14, 2023
July 16, 2023
June 12, 2023
May 19, 2023
May 5, 2023
November 21, 2022
November 8, 2022
November 3, 2022
October 27, 2022
October 13, 2022
October 5, 2022
September 23, 2022
August 4, 2022
July 28, 2022
July 14, 2022