Recent Large Language Model
Recent research on large language models (LLMs) centers on improving their capabilities in handling long contexts, multilingual support, and complex reasoning tasks, while also addressing limitations in efficiency, bias, and uncertainty quantification. Current efforts focus on novel architectures like Mamba, enhanced Mixture of Experts models, and improved training methods such as self-contrast learning and fine-grained reward systems. These advancements are crucial for expanding the practical applications of LLMs across diverse fields, from biomedical research and public health interventions to improving the reliability of AI-assisted tools and mitigating the risks associated with misinformation.
Papers
June 11, 2024
May 27, 2024
May 23, 2024
May 8, 2024
April 22, 2024
April 3, 2024
March 31, 2024
March 23, 2024
March 2, 2024
February 26, 2024
February 17, 2024
February 13, 2024
February 6, 2024
February 5, 2024
January 12, 2024
December 25, 2023
November 15, 2023
November 9, 2023
November 6, 2023