Instruction Following
Instruction following in large language models (LLMs) focuses on enhancing their ability to accurately and reliably execute diverse instructions, a crucial step towards building truly general-purpose AI. Current research emphasizes improving generalization by diversifying training data across semantic domains and optimizing data sampling strategies, often employing techniques like clustering and iterative refinement. These advancements are significant because robust instruction following is essential for safe and effective deployment of LLMs in various applications, ranging from assisting researchers in navigating scientific literature to automating complex tasks in manufacturing. Furthermore, research is actively exploring methods to improve the reliability and robustness of instruction following, including mitigating catastrophic forgetting and addressing vulnerabilities to adversarial attacks.
Papers
Stronger Models are NOT Stronger Teachers for Instruction Tuning
Zhangchen Xu, Fengqing Jiang, Luyao Niu, Bill Yuchen Lin, Radha Poovendran
LIFBench: Evaluating the Instruction Following Performance and Stability of Large Language Models in Long-Context Scenarios
Xiaodong Wu, Minhao Wang, Yichen Liu, Xiaoming Shi, He Yan, Xiangju Lu, Junmin Zhu, Wei Zhang