Open Weight Model
Open-weight models are large language models (LLMs) whose parameters are publicly available, fostering collaboration and reproducibility in AI research. Current research focuses on improving their performance, particularly through novel training datasets (including multimodal data) and architectural innovations like hybrid transformer-SSM designs, aiming to match or surpass the capabilities of closed-source models. This open approach is significant because it democratizes access to advanced LLMs, enabling broader participation in AI development and facilitating research into their capabilities and limitations across diverse domains and languages, including specialized scientific fields like astronomy.
Papers
September 25, 2024
September 23, 2024
August 15, 2024
July 15, 2024
June 6, 2024
May 27, 2024
May 26, 2024
March 29, 2024
February 26, 2024