Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
How much can we forget about Data Contamination?
Sebastian Bordt, Suraj Srinivas, Valentyn Boreiko, Ulrike von Luxburg
Showing LLM-Generated Code Selectively Based on Confidence of LLMs
Jia Li, Yuqi Zhu, Yongmin Li, Ge Li, Zhi Jin
ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering
Huayang Li, Pat Verga, Priyanka Sen, Bowen Yang, Vijay Viswanathan, Patrick Lewis, Taro Watanabe, Yixuan Su
Frame-Voyager: Learning to Query Frames for Video Large Language Models
Sicheng Yu, Chengkai Jin, Huanyu Wang, Zhenghao Chen, Sheng Jin, Zhongrong Zuo, Xiaolei Xu, Zhenbang Sun, Bingni Zhang, Jiawei Wu, Hao Zhang, Qianru Sun
PersoBench: Benchmarking Personalized Response Generation in Large Language Models
Saleh Afzoon, Usman Naseem, Amin Beheshti, Zahra Jamali
EXAQ: Exponent Aware Quantization For LLMs Acceleration
Moran Shkolnik, Maxim Fishman, Brian Chmiel, Hilla Ben-Yaacov, Ron Banner, Kfir Yehuda Levy
Generating bilingual example sentences with large language models as lexicography assistants
Raphael Merx, Ekaterina Vylomova, Kemal Kurniawan
Autoregressive Large Language Models are Computationally Universal
Dale Schuurmans, Hanjun Dai, Francesco Zanini
Margin Matching Preference Optimization: Enhanced Model Alignment with Granular Feedback
Kyuyoung Kim, Ah Jeong Seo, Hao Liu, Jinwoo Shin, Kimin Lee
How Do Large Language Models Understand Graph Patterns? A Benchmark for Graph Pattern Comprehension
Xinnan Dai, Haohao Qu, Yifen Shen, Bohang Zhang, Qihao Wen, Wenqi Fan, Dongsheng Li, Jiliang Tang, Caihua Shan
Can LLMs Generate Diverse Molecules? Towards Alignment with Structural Diversity
Hyosoon Jang, Yunhui Jang, Jaehyung Kim, Sungsoo Ahn
SAG: Style-Aligned Article Generation via Model Collaboration
Chenning Xu, Fangxun Shu, Dian Jin, Jinghao Wei, Hao Jiang
Deliberate Reasoning for LLMs as Structure-aware Planning with Accurate World Model
Siheng Xiong, Ali Payani, Yuan Yang, Faramarz Fekri
Remaining Useful Life Prediction: A Study on Multidimensional Industrial Signal Processing and Efficient Transfer Learning Based on Large Language Models
Yan Chen, Cheng Liu
Searching for Best Practices in Medical Transcription with Large Language Model
Jiafeng Li, Yanda Mu
On Unsupervised Prompt Learning for Classification with Black-box Language Models
Zhen-Yu Zhang, Jiandong Zhang, Huaxiu Yao, Gang Niu, Masashi Sugiyama
RIPPLECOT: Amplifying Ripple Effect of Knowledge Editing in Language Models via Chain-of-Thought In-Context Learning
Zihao Zhao, Yuchen Yang, Yijiang Li, Yinzhi Cao
X-ALMA: Plug & Play Modules and Adaptive Rejection for Quality Translation at Scale
Haoran Xu, Kenton Murray, Philipp Koehn, Hieu Hoang, Akiko Eriguchi, Huda Khayrallah
Image First or Text First? Optimising the Sequencing of Modalities in Large Language Model Prompting and Reasoning Tasks
Grant Wardle, Teo Susnjak