General Language Model
General language models (GLMs) are large neural networks trained on massive text datasets to understand and generate human language. Current research focuses on adapting GLMs to specific domains (e.g., medicine, cybersecurity) through techniques like fine-tuning and knowledge distillation, while also addressing challenges such as maintaining long-context understanding and mitigating biases. These advancements are significantly impacting various fields, enabling improved natural language processing tasks in diverse applications and offering new tools for analyzing specialized corpora.
Papers
September 23, 2024
July 16, 2024
June 14, 2024
May 28, 2024
February 10, 2024
November 16, 2023
November 14, 2023
October 19, 2023
October 16, 2023
August 28, 2023
August 7, 2023
June 14, 2023
June 13, 2023
May 20, 2023
February 13, 2023
December 14, 2022