Instruction Following
Instruction following in large language models (LLMs) focuses on enhancing their ability to accurately and reliably execute diverse instructions, a crucial step towards building truly general-purpose AI. Current research emphasizes improving generalization by diversifying training data across semantic domains and optimizing data sampling strategies, often employing techniques like clustering and iterative refinement. These advancements are significant because robust instruction following is essential for safe and effective deployment of LLMs in various applications, ranging from assisting researchers in navigating scientific literature to automating complex tasks in manufacturing. Furthermore, research is actively exploring methods to improve the reliability and robustness of instruction following, including mitigating catastrophic forgetting and addressing vulnerabilities to adversarial attacks.
Papers
Benchmarking Complex Instruction-Following with Multiple Constraints Composition
Bosi Wen, Pei Ke, Xiaotao Gu, Lindong Wu, Hao Huang, Jinfeng Zhou, Wenchuang Li, Binxin Hu, Wendy Gao, Jiaxin Xu, Yiming Liu, Jie Tang, Hongning Wang, Minlie Huang
Diverse and Fine-Grained Instruction-Following Ability Exploration with Synthetic Data
Zihui Gu, Xingwu Sun, Fengzong Lian, Zhanhui Kang, Cheng-Zhong Xu, Ju Fan
The Comparative Trap: Pairwise Comparisons Amplifies Biased Preferences of LLM Evaluators
Hawon Jeong, ChaeHun Park, Jimin Hong, Hojoon Lee, Jaegul Choo
Interpretable Catastrophic Forgetting of Large Language Model Fine-tuning via Instruction Vector
Gangwei Jiang, Caigao Jiang, Zhaoyi Li, Siqiao Xue, Jun Zhou, Linqi Song, Defu Lian, Ying Wei