Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Resource Allocation for Stable LLM Training in Mobile Edge Computing
Chang Liu, Jun Zhao
Reference Trustable Decoding: A Training-Free Augmentation Paradigm for Large Language Models
Luohe Shi, Yao Yao, Zuchao Li, Lefei Zhang, Hai Zhao
Classification of Radiological Text in Small and Imbalanced Datasets in a Non-English Language
Vincent Beliveau, Helene Kaas, Martin Prener, Claes N. Ladefoged, Desmond Elliott, Gitte M. Knudsen, Lars H. Pinborg, Melanie Ganz
Beyond Scores: A Modular RAG-Based System for Automatic Short Answer Scoring with Feedback
Menna Fateen, Bo Wang, Tsunenori Mine
Developing Instruction-Following Speech Language Model Without Speech Instruction-Tuning Data
Ke-Han Lu, Zhehuai Chen, Szu-Wei Fu, Chao-Han Huck Yang, Jagadeesh Balam, Boris Ginsburg, Yu-Chiang Frank Wang, Hung-yi Lee
Do Influence Functions Work on Large Language Models?
Zhe Li, Wei Zhao, Yige Li, Jun Sun
FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows"
Yifei Ming, Senthil Purushwalkam, Shrey Pandit, Zixuan Ke, Xuan-Phi Nguyen, Caiming Xiong, Shafiq Joty
Enhancing High-order Interaction Awareness in LLM-based Recommender Model
Xinfeng Wang, Jin Cui, Fumiyo Fukumoto, Yoshimi Suzuki
RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
Shuhao Chen, Weisen Jiang, Baijiong Lin, James T. Kwok, Yu Zhang
Neurosymbolic AI approach to Attribution in Large Language Models
Deepa Tilwani, Revathy Venkataramanan, Amit P. Sheth
Calibrating Language Models with Adaptive Temperature Scaling
Johnathan Xie, Annie S. Chen, Yoonho Lee, Eric Mitchell, Chelsea Finn
Transforming Hidden States into Binary Semantic Features
Tomáš Musil, David Mareček
Balancing Cost and Effectiveness of Synthetic Data Generation Strategies for LLMs
Yung-Chieh Chan, George Pu, Apaar Shanker, Parth Suresh, Penn Jenks, John Heyer, Sam Denton
AstroMLab 2: AstroLLaMA-2-70B Model and Benchmarking Specialised LLMs for Astronomy
Rui Pan, Tuan Dung Nguyen, Hardik Arora, Alberto Accomazzi, Tirthankar Ghosal, Yuan-Sen Ting
PEAR: Position-Embedding-Agnostic Attention Re-weighting Enhances Retrieval-Augmented Generation with Zero Inference Overhead
Tao Tan, Yining Qian, Ang Lv, Hongzhan Lin, Songhao Wu, Yongbo Wang, Feng Wang, Jingtong Wu, Xin Lu, Rui Yan
2D-TPE: Two-Dimensional Positional Encoding Enhances Table Understanding for Large Language Models
Jia-Nan Li, Jian Guan, Wei Wu, Zhengtao Yu, Rui Yan
Learning Attentional Mixture of LoRAs for Language Model Continual Learning
Jialin Liu, Jianhua Wu, Jie Liu, Yutai Duan
Hyper-Connections
Defa Zhu, Hongzhi Huang, Zihao Huang, Yutao Zeng, Yunyao Mao, Banggu Wu, Qiyang Min, Xun Zhou
LANDeRMT: Detecting and Routing Language-Aware Neurons for Selectively Finetuning LLMs to Machine Translation
Shaolin Zhu, Leiyu Pan, Bo Li, Deyi Xiong
MedHalu: Hallucinations in Responses to Healthcare Queries by Large Language Models
Vibhor Agarwal, Yiqiao Jin, Mohit Chandra, Munmun De Choudhury, Srijan Kumar, Nishanth Sastry