Open Source Large Language Model
Open-source large language models (LLMs) aim to provide accessible and customizable alternatives to proprietary models, fostering research and development while addressing concerns about data privacy and vendor lock-in. Current research focuses on adapting these models to specific languages and domains (e.g., Romanian, medicine, finance), improving their reasoning capabilities through techniques like retrieval-augmented generation and mixture-of-experts architectures, and optimizing their deployment efficiency on various hardware. This burgeoning field significantly impacts both the scientific community, by enabling broader participation in LLM research, and practical applications, offering cost-effective and adaptable solutions for diverse tasks ranging from question answering to code generation.
Papers
WikiContradict: A Benchmark for Evaluating LLMs on Real-World Knowledge Conflicts from Wikipedia
Yufang Hou, Alessandra Pascale, Javier Carnerero-Cano, Tigran Tchrakian, Radu Marinescu, Elizabeth Daly, Inkit Padhi, Prasanna Sattigeri
Benchmarking Open-Source Language Models for Efficient Question Answering in Industrial Applications
Mahaman Sanoussi Yahaya Alassan, Jessica López Espejel, Merieme Bouhandi, Walid Dahhane, El Hassane Ettifouri
LLMs Are Zero-Shot Context-Aware Simultaneous Translators
Roman Koshkin, Katsuhito Sudoh, Satoshi Nakamura
Efficient Prompting for LLM-based Generative Internet of Things
Bin Xiao, Burak Kantarci, Jiawen Kang, Dusit Niyato, Mohsen Guizani
A Survey on Large Language Models from General Purpose to Medical Applications: Datasets, Methodologies, and Evaluations
Jinqiang Wang, Huansheng Ning, Yi Peng, Qikai Wei, Daniel Tesfai, Wenwei Mao, Tao Zhu, Runhe Huang
Representation noising can prevent harmful fine-tuning on LLMs
Domenic Rosati, Jan Wehner, Kai Williams, Łukasz Bartoszcze, David Atanasov, Robie Gonzales, Subhabrata Majumdar, Carsten Maple, Hassan Sajjad, Frank Rudzicz
MoGU: A Framework for Enhancing Safety of Open-Sourced LLMs While Preserving Their Usability
Yanrui Du, Sendong Zhao, Danyang Zhao, Ming Ma, Yuhan Chen, Liangyu Huo, Qing Yang, Dongliang Xu, Bing Qin