Instruction Generation
Instruction generation focuses on automatically creating high-quality instructions for various tasks, primarily to improve the performance of large language models (LLMs) and other AI agents. Current research emphasizes developing robust methods for generating diverse and complex instructions, often employing techniques like adversarial training, evolutionary algorithms, and chain-of-thought prompting within transformer-based architectures. This field is crucial for advancing AI capabilities across numerous domains, from robotics and virtual navigation to question answering and multimodal learning, by providing more effective training data and enabling more natural human-computer interaction.
Papers
Bootstrapping Language-Guided Navigation Learning with Self-Refining Data Flywheel
Zun Wang, Jialu Li, Yicong Hong, Songze Li, Kunchang Li, Shoubin Yu, Yi Wang, Yu Qiao, Yali Wang, Mohit Bansal, Limin Wang
Template Matters: Understanding the Role of Instruction Templates in Multimodal Language Model Evaluation and Training
Shijian Wang, Linxin Song, Jieyu Zhang, Ryotaro Shimizu, Ao Luo, Li Yao, Cunjian Chen, Julian McAuley, Hanqian Wu