Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Improving LLM Group Fairness on Tabular Data via In-Context Learning
Valeriia Cherepanova, Chia-Jung Lee, Nil-Jana Akpinar, Riccardo Fogliato, Martin Andres Bertran, Michael Kearns, James Zou
Give me Some Hard Questions: Synthetic Data Generation for Clinical QA
Fan Bai, Keith Harrigian, Joel Stremmel, Hamid Hassanzadeh, Ardavan Saeedi, Mark Dredze
EgoPlan-Bench2: A Benchmark for Multimodal Large Language Model Planning in Real-World Scenarios
Lu Qiu, Yuying Ge, Yi Chen, Yixiao Ge, Ying Shan, Xihui Liu
Understanding Hidden Computations in Chain-of-Thought Reasoning
Aryasomayajula Ram Bharadwaj
Retrieval-Augmented Machine Translation with Unstructured Knowledge
Jiaan Wang, Fandong Meng, Yingxue Zhang, Jie Zhou
Liquid: Language Models are Scalable Multi-modal Generators
Junfeng Wu, Yi Jiang, Chuofan Ma, Yuliang Liu, Hengshuang Zhao, Zehuan Yuan, Song Bai, Xiang Bai
The Hyperfitting Phenomenon: Sharpening and Stabilizing LLMs for Open-Ended Text Generation
Fredrik Carlsson, Fangyu Liu, Daniel Ward, Murathan Kurfali, Joakim Nivre
FlashSloth: Lightning Multimodal Large Language Models via Embedded Visual Compression
Bo Tong, Bokai Lai, Yiyi Zhou, Gen Luo, Yunhang Shen, Ke Li, Xiaoshuai Sun, Rongrong Ji
Densing Law of LLMs
Chaojun Xiao, Jie Cai, Weilin Zhao, Guoyang Zeng, Xu Han, Zhiyuan Liu, Maosong Sun
ALMA: Alignment with Minimal Annotation
Michihiro Yasunaga, Leonid Shamis, Chunting Zhou, Andrew Cohen, Jason Weston, Luke Zettlemoyer, Marjan Ghazvininejad
Evolutionary Pre-Prompt Optimization for Mathematical Reasoning
Mathurin Videau, Alessandro Leite, Marc Schoenauer, Olivier Teytaud
AL-QASIDA: Analyzing LLM Quality and Accuracy Systematically in Dialectal Arabic
Nathaniel R. Robinson, Shahd Abdelmoneim, Kelly Marchisio, Sebastian Ruder
Leveraging Large Language Models to Generate Course-specific Semantically Annotated Learning Objects
Dominic Lohr, Marc Berges, Abhishek Chugh, Michael Kohlhase, Dennis Müller
Monet: Mixture of Monosemantic Experts for Transformers
Jungwoo Park, Young Jin Ahn, Kee-Eung Kim, Jaewoo Kang
Pre-train, Align, and Disentangle: Empowering Sequential Recommendation with Large Language Models
Yuhao Wang, Junwei Pan, Xiangyu Zhao, Pengyue Jia, Wanyu Wang, Yuan Wang, Yue Liu, Dapeng Liu, Jie Jiang
Guidance is All You Need: Temperature-Guided Reasoning in Large Language Models
Eyad Gomaa, Gomaa Salah
Practical Considerations for Agentic LLM Systems
Chris Sypherd, Vaishak Belle
LossAgent: Towards Any Optimization Objectives for Image Processing with LLM Agents
Bingchen Li, Xin Li, Yiting Lu, Zhibo Chen
MTMT: Consolidating Multiple Thinking Modes to Form a Thought Tree for Strengthening LLM
Changcheng Li, Xiangyu Wang, Qiuju Chen, Xiren Zhou, Huanhuan Chen