Open Source Language Model
Open-source language models (LLMs) aim to democratize access to and research on powerful language AI by making model weights, training data, and code publicly available. Current research focuses on improving model performance through techniques like fine-tuning with specialized datasets (e.g., for medical summarization or mathematical reasoning), model merging to combine strengths, and addressing security vulnerabilities like backdoor attacks. This open approach fosters collaboration, accelerates innovation, and enables applications across diverse fields, including healthcare, education, and scientific research, while also raising important questions about data bias, model safety, and ethical considerations.
Papers
October 24, 2024
October 23, 2024
October 20, 2024
October 19, 2024
October 17, 2024
July 3, 2024
June 19, 2024
June 10, 2024
June 5, 2024
May 25, 2024
May 22, 2024
May 8, 2024
May 2, 2024
April 26, 2024
April 25, 2024
April 15, 2024
March 20, 2024
March 5, 2024