Instruction Tuning
Instruction tuning refines large language models (LLMs) by training them on datasets of instructions and desired responses, improving their ability to follow diverse commands and generate helpful outputs. Current research emphasizes improving data quality and diversity through techniques like data partitioning, synthetic data generation, and novel prompting strategies, often applied to various model architectures including LLMs and multimodal models. This area is significant because it directly addresses the limitations of pre-trained LLMs, leading to safer, more reliable, and more useful AI systems across numerous applications, from chatbots to specialized tools for medical diagnosis and remote sensing.
Papers
Automatic Evaluation of Generative Models with Instruction Tuning
Shuhaib Mehri, Vered Shwartz
BioInstruct: Instruction Tuning of Large Language Models for Biomedical Natural Language Processing
Hieu Tran, Zhichao Yang, Zonghai Yao, Hong Yu
Dynamics of Instruction Tuning: Each Ability of Large Language Models Has Its Own Growth Pace
Chiyu Song, Zhanchao Zhou, Jianhao Yan, Yuejiao Fei, Zhenzhong Lan, Yue Zhang
Specialist or Generalist? Instruction Tuning for Specific NLP Tasks
Chufan Shi, Yixuan Su, Cheng Yang, Yujiu Yang, Deng Cai
AlpaCare:Instruction-tuned Large Language Models for Medical Application
Xinlu Zhang, Chenxin Tian, Xianjun Yang, Lichang Chen, Zekun Li, Linda Ruth Petzold
CITB: A Benchmark for Continual Instruction Tuning
Zihan Zhang, Meng Fang, Ling Chen, Mohammad-Reza Namazi-Rad