Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
SimulBench: Evaluating Language Models with Creative Simulation Tasks
Qi Jia, Xiang Yue, Tianyu Zheng, Jie Huang, Bill Yuchen Lin
When Less Is Not More: Large Language Models Normalize Less-Frequent Terms with Lower Accuracy
Daniel B. Hier, Thanh Son Do, Tayo Obafemi-Ajayi
Contextualization of ASR with LLM using phonetic retrieval-based augmentation
Zhihong Lei, Xingyu Na, Mingbin Xu, Ernest Pusateri, Christophe Van Gysel, Yuanyuan Zhang, Shiyi Han, Zhen Huang
AdaCAD: Adaptively Decoding to Balance Conflicts between Contextual and Parametric Knowledge
Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal
Think Together and Work Better: Combining Humans' and LLMs' Think-Aloud Outcomes for Effective Text Evaluation
SeongYeub Chu, JongWoo Kim, MunYong Yi
MEDIC: Towards a Comprehensive Framework for Evaluating LLMs in Clinical Applications
Praveen K Kanithi, Clément Christophe, Marco AF Pimentel, Tathagata Raha, Nada Saadi, Hamza Javed, Svetlana Maslenkova, Nasir Hayat, Ronnie Rajan, Shadab Khan
Alignment of Diffusion Models: Fundamentals, Challenges, and Future
Buhua Liu, Shitong Shao, Bao Li, Lichen Bai, Haoyi Xiong, James Kwok, Sumi Helal, Zeke Xie
LLM-based feature generation from text for interpretable machine learning
Vojtěch Balek, Lukáš Sýkora, Vilém Sklenák, Tomáš Kliegr
Reranking Laws for Language Generation: A Communication-Theoretic Perspective
António Farinhas, Haau-Sing Li, André F. T. Martins
Cross-Refine: Improving Natural Language Explanation Generation by Learning in Tandem
Qianli Wang, Tatiana Anikina, Nils Feldhus, Simon Ostermann, Sebastian Möller, Vera Schmitt
Ontology-Free General-Domain Knowledge Graph-to-Text Generation Dataset Synthesis using Large Language Model
Daehee Kim, Deokhyung Kang, Sangwon Ryu, Gary Geunbae Lee
Understanding Knowledge Drift in LLMs through Misinformation
Alina Fastowski, Gjergji Kasneci
Native vs Non-Native Language Prompting: A Comparative Analysis
Mohamed Bayan Kmainasi, Rakif Khan, Ali Ezzat Shahroor, Boushra Bendou, Maram Hasanain, Firoj Alam
Beyond IID: Optimizing Instruction Learning from the Perspective of Instruction Interaction and Dependency
Hanyu Zhao, Li Du, Yiming Ju, Chengwei Wu, Tengfei Pan
Representation Tuning
Christopher M. Ackerman
A Simplified Retriever to Improve Accuracy of Phenotype Normalizations by Large Language Models
Daniel B. Hier, Thanh Son Do, Tayo Obafemi-Ajayi
AdaPPA: Adaptive Position Pre-Fill Jailbreak Attack Approach Targeting LLMs
Lijia Lv, Weigang Zhang, Xuehai Tang, Jie Wen, Feng Liu, Jizhong Han, Songlin Hu
Semi-Supervised Reward Modeling via Iterative Self-Training
Yifei He, Haoxiang Wang, Ziyan Jiang, Alexandros Papangelis, Han Zhao
MoWE-Audio: Multitask AudioLLMs with Mixture of Weak Encoders
Wenyu Zhang, Shuo Sun, Bin Wang, Xunlong Zou, Zhuohan Liu, Yingxu He, Geyu Lin, Nancy F. Chen, Ai Ti Aw
Beyond designer's knowledge: Generating materials design hypotheses via large language models
Quanliang Liu, Maciej P. Polak, So Yeon Kim, MD Al Amin Shuvo, Hrishikesh Shridhar Deodhar, Jeongsoo Han, Dane Morgan, Hyunseok Oh