Human Feedback
Human feedback is crucial for aligning artificial intelligence models, particularly large language models, with human preferences and values. Current research focuses on improving the efficiency and reliability of incorporating human feedback into reinforcement learning frameworks, exploring techniques like macro actions, active learning, and reward model optimization to address challenges such as the cost and subjectivity of human judgments. This work is significant because it directly impacts the safety, trustworthiness, and overall effectiveness of AI systems across diverse applications, from autonomous driving to educational assessment. The development of more robust and efficient methods for integrating human feedback is a key area of ongoing investigation.
Papers
Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback
Jiakang Yuan, Xiangchao Yan, Botian Shi, Tao Chen, Wanli Ouyang, Bo Zhang, Lei Bai, Yu Qiao, Bowen Zhou
Effects of Robot Competency and Motion Legibility on Human Correction Feedback
Shuangge Wang, Anjiabei Wang, Sofiya Goncharova, Brian Scassellati, Tesca Fitzgerald
Test-time Correction with Human Feedback: An Online 3D Detection System via Visual Prompting
Zetong Yang, Hanxue Zhang, Yanan Sun, Li Chen, Fei Xia, Fatma Guney, Hongyang Li
Enhancing Relation Extraction via Supervised Rationale Verification and Feedback
Yongqi Li, Xin Miao, Shen Zhou, Mayi Xu, Yuyang Ren, Tieyun Qian