Open Source Model
Open-source large language models (LLMs) aim to democratize access to powerful AI by providing freely available model weights, code, and sometimes even training data. Current research focuses on improving the performance and safety of these models, including developing novel training techniques, exploring efficient model compression methods like pruning and merging, and establishing robust benchmarks for evaluating trustworthiness, bias, and safety. This open approach fosters collaboration, accelerates innovation, and addresses concerns about proprietary model limitations, particularly regarding data privacy and accessibility for researchers and developers in various fields.
Papers
ProFLingo: A Fingerprinting-based Intellectual Property Protection Scheme for Large Language Models
Heng Jin, Chaoyu Zhang, Shanghao Shi, Wenjing Lou, Y. Thomas Hou
Aloe: A Family of Fine-tuned Open Healthcare LLMs
Ashwin Kumar Gururajan, Enrique Lopez-Cuena, Jordi Bayarri-Planas, Adrian Tormos, Daniel Hinjos, Pablo Bernabeu-Perez, Anna Arias-Duart, Pablo Agustin Martin-Torres, Lucia Urcelay-Ganzabal, Marta Gonzalez-Mallo, Sergio Alvarez-Napagao, Eduard Ayguadé-Parra, Ulises Cortés Dario Garcia-Gasulla